File size: 3,020 Bytes
0e930f7 ce46fb8 c6fb9a5 9b070c9 5950949 37dcc47 1bcadaa 22b66b6 52608a4 c2e9c27 ea0c1b0 d785e14 35d8035 edaf76b f6458ff f5baaec 061ecb1 48ef7a3 59cb78a 7bc67dc f9a96ca 75106a1 43bc10a 40c3b1b 4bc24c5 c232b18 c643afa e48c0b1 f6087f3 2e31265 18b69b7 104d88f 967dc1a 31eeb03 f77d678 e8b9dab 73ed8d3 b1bf76f 499d2c2 012839b 8c36c93 b5ced94 0093543 08b7287 d0cddc6 d7d2794 5ff58c7 2715da0 ce0afad 0d5b9c8 e1d2b47 74d3b1d 7d0c4fe 438bcd1 f941247 2357641 4e1b70a eedf4e8 1d07ac8 47601af 8eb63c7 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 |
Started at: 10:20:56 norbert2, 0.001, 320 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.3527084864675999 - MAE: 0.462953214037108 Validation loss : 0.26738621081624714 - MAE: 0.4117436401566978 Epoch: 1 Training loss: 0.21354530677199363 - MAE: 0.35412568088618607 Validation loss : 0.18583983182907104 - MAE: 0.3299781625243953 Epoch: 2 Training loss: 0.18886808156967164 - MAE: 0.330421134278654 Validation loss : 0.1822844914027623 - MAE: 0.33056134626981226 Epoch: 3 Training loss: 0.18264293819665908 - MAE: 0.3242538405037828 Validation loss : 0.1772120531116213 - MAE: 0.3251080552524876 Epoch: 4 Training loss: 0.17669695615768433 - MAE: 0.3190416206929785 Validation loss : 0.1738429559128625 - MAE: 0.322266401087556 Epoch: 5 Training loss: 0.17284812778234482 - MAE: 0.3136576419048662 Validation loss : 0.17294207853930338 - MAE: 0.32187514594462957 Epoch: 6 Training loss: 0.16915602013468742 - MAE: 0.3109971127550157 Validation loss : 0.16980814508029393 - MAE: 0.3186160670865416 Epoch: 7 Training loss: 0.16845666095614434 - MAE: 0.3089849145668241 Validation loss : 0.16714746824332646 - MAE: 0.3155506548087574 Epoch: 8 Training loss: 0.16868472024798392 - MAE: 0.30949657276040055 Validation loss : 0.16395465178149088 - MAE: 0.3112663647803296 Epoch: 9 Training loss: 0.16725652366876603 - MAE: 0.3066475873241304 Validation loss : 0.1607010385819844 - MAE: 0.3061224215087366 Epoch: 10 Training loss: 0.16232142224907875 - MAE: 0.30304105641340856 Validation loss : 0.1601495338337762 - MAE: 0.30464326769452593 Epoch: 11 Training loss: 0.1600735753774643 - MAE: 0.3024141941842313 Validation loss : 0.16005327233246394 - MAE: 0.30402928496694465 Epoch: 12 Training loss: 0.1590587668120861 - MAE: 0.30053309807554257 Validation loss : 0.16130509546824864 - MAE: 0.30322365490516023 Epoch: 13 Training loss: 0.1575142566114664 - MAE: 0.3000214052697603 Validation loss : 0.16007075778075627 - MAE: 0.3021056102178289 Epoch: 14 Training loss: 0.15866125449538232 - MAE: 0.2997664917473247 Validation loss : 0.15920681825705937 - MAE: 0.30173646847418345 Epoch: 15 Training loss: 0.15266035348176957 - MAE: 0.2944185943451775 Validation loss : 0.15711544241224015 - MAE: 0.3000104381366772 Epoch: 16 Training loss: 0.15278259553015233 - MAE: 0.29361855853063534 Validation loss : 0.1574518233537674 - MAE: 0.29962551415919003 Epoch: 17 Training loss: 0.15116168446838857 - MAE: 0.2937445197592836 Validation loss : 0.1565468375171934 - MAE: 0.29831129957449803 Epoch: 18 Training loss: 0.1503951381891966 - MAE: 0.2918888435388691 Validation loss : 0.15591060050896235 - MAE: 0.2980704319089749 Epoch: 19 |