File size: 4,819 Bytes
34f0ece 6222105 75a0e74 64a9570 546d517 68b0f20 0939508 fa6ba90 3819732 0618aa1 9f90d1d 7bb073c 5abda92 29bd34a 24e668c 835b927 13f908d 99549f4 faea999 0eda6c3 34eddc9 22c61bc a388479 7f8bb19 b14a98f 795590a 89a4235 d8571b3 03378a7 dd8464b 435981b aa21c72 914c67e 0c36e41 1c32fb7 685e975 0a29e9b 39603f5 44fc5bd 36c07c6 57e9c2d e5787b5 bd1456c ba70980 d2a29ae cc65b11 1c6ed4f 7a2aeed 5eb6ae4 352f378 32dbd56 c6b7c40 5a1c670 9f633b3 734ab14 a8833f4 64288a9 705ba6a c38e3e5 ce59bd9 c50386c 1a7a4b5 3239c28 7036e56 3510d16 8dd4bad f06365a 6b96172 e2ba096 007fc8c cf0202e f95af6f 4cbd93e de0aebd 7ace8ad 4c5e61c e15afd5 ea181e1 66b262a 132f5c1 6b361ed ccab03b cd0ecb4 3f7c7d4 d2fc893 71e9ad8 0132a7e e1bc946 ea4acff e487d9f 52d867c 90398dd 061f920 4881735 f3f0853 00b62f9 f68044a 0dbe926 377a5fd 319f4c1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 |
Started at: 13:54:54 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '44815f7e109b53547cccdf3c6847f4c28b989816'}, {}) Epoch: 0 Training loss: 0.4353753740065976 - MSE: 0.4631018830797368 Validation loss : 0.14201142988167703 - MSE: 0.292727476662094 Epoch: 1 Training loss: 0.16082680688092582 - MSE: 0.3039855683631176 Validation loss : 0.14794749207794666 - MSE: 0.29952175843163786 Epoch: 2 Training loss: 0.15888278323568797 - MSE: 0.30232633269739506 Validation loss : 0.12977031315676868 - MSE: 0.2779232412353849 Epoch: 3 Training loss: 0.1516539692878723 - MSE: 0.29570308847882737 Validation loss : 0.1342183407396078 - MSE: 0.28315487688041685 Epoch: 4 Training loss: 0.14969512843771984 - MSE: 0.29376316316788137 Validation loss : 0.13651572703383863 - MSE: 0.286356273434194 Epoch: 5 Training loss: 0.14825457373732015 - MSE: 0.29236288646470276 Validation loss : 0.13325480837374926 - MSE: 0.28235259219172804 Epoch: 6 Training loss: 0.14587709978222846 - MSE: 0.29145181360233374 Validation loss : 0.14294268772937357 - MSE: 0.29422283310077546 Epoch: 7 Training loss: 0.148439571535901 - MSE: 0.29440375310883926 Validation loss : 0.14034398854710162 - MSE: 0.2907182476260459 Epoch: 8 Training loss: 0.14813685872052845 - MSE: 0.2931949322407332 Validation loss : 0.12979571800678968 - MSE: 0.2782447249405777 Epoch: 9 Training loss: 0.14627560220266644 - MSE: 0.2917503264555033 Validation loss : 0.1504771877080202 - MSE: 0.30260855387996344 Epoch: 10 Training loss: 0.14390636309981347 - MSE: 0.29031416543032734 Validation loss : 0.13953579450026155 - MSE: 0.2890264796219526 Epoch: 11 Training loss: 0.14922026603629715 - MSE: 0.292561392520073 Validation loss : 0.1444824270438403 - MSE: 0.2944073201865649 Epoch: 12 Training loss: 0.14153452319534202 - MSE: 0.285389527479883 Validation loss : 0.1263227758463472 - MSE: 0.27377457326610966 Epoch: 13 Training loss: 0.14998281272618394 - MSE: 0.2927288895743673 Validation loss : 0.12917391466908157 - MSE: 0.27674699036970196 Epoch: 14 Training loss: 0.1430773071160442 - MSE: 0.2892675879716682 Validation loss : 0.14142304356209934 - MSE: 0.2941539923554046 Epoch: 15 Training loss: 0.15678064901577798 - MSE: 0.3018205089469717 Validation loss : 0.12917302176356316 - MSE: 0.27513299298107086 Epoch: 16 Training loss: 0.1494230675854181 - MSE: 0.2951704079181553 Validation loss : 0.1280019199475646 - MSE: 0.2731706016706994 Epoch: 17 Training loss: 0.14316309780666703 - MSE: 0.28891571660292187 Validation loss : 0.12425365822855383 - MSE: 0.27112690135936646 Epoch: 18 Training loss: 0.1434906577593402 - MSE: 0.28606728956978134 Validation loss : 0.15053234971128404 - MSE: 0.30457389003231583 Epoch: 19 Training loss: 0.14480229098545877 - MSE: 0.2907542990574569 Validation loss : 0.1469436811748892 - MSE: 0.3000679240572026 Epoch: 20 Training loss: 0.1462563581372562 - MSE: 0.2903288957751347 Validation loss : 0.16302492283284664 - MSE: 0.3174909429867512 Epoch: 21 Training loss: 0.14268505094867004 - MSE: 0.28808178734771134 Validation loss : 0.12645738548599184 - MSE: 0.2734583813939935 Epoch: 22 Training loss: 0.1431882705735533 - MSE: 0.29013400485776225 Validation loss : 0.14717783266678452 - MSE: 0.3002806941290146 Epoch: 23 Training loss: 0.1427597692922542 - MSE: 0.2874371575758064 Validation loss : 0.12963575683534145 - MSE: 0.2767674714336863 Epoch: 24 Training loss: 0.14602290623282133 - MSE: 0.29132534981254526 Validation loss : 0.15562276146374643 - MSE: 0.30841544617533145 Epoch: 25 Training loss: 0.14057497789985254 - MSE: 0.2866385142998687 Validation loss : 0.12258221080992371 - MSE: 0.2676300164089298 Epoch: 26 Training loss: 0.13991084024310113 - MSE: 0.2862323305934003 Validation loss : 0.14549001771956682 - MSE: 0.29673826599128006 Epoch: 27 Training loss: 0.13917529892764594 - MSE: 0.2846490242675938 Validation loss : 0.1338302984368056 - MSE: 0.2824601715906283 Epoch: 28 Training loss: 0.1368718992330526 - MSE: 0.28348191048819527 Validation loss : 0.13304794486612082 - MSE: 0.28124178885582296 Epoch: 29 Training loss: 0.137821614467784 - MSE: 0.28195915514916314 Validation loss : 0.1356566457543522 - MSE: 0.2844219791945761 Epoch: 30 Training loss: 0.1361325614154339 - MSE: 0.2826986827887364 Validation loss : 0.13432824215851724 - MSE: 0.28257486897155104 Epoch: 31 Training loss: 0.13651365814240354 - MSE: 0.28391514279453545 Validation loss : 0.131905903108418 - MSE: 0.2781696847882813 Epoch: 32 Training loss: 0.13995016409378302 - MSE: 0.28707461995686306 |