File size: 2,729 Bytes
a359b2c 74ff63e b6f2008 7a00688 6ebe937 44b78fe 845a0b9 f777ecd 05eddfb 63ce967 f768d92 b978586 2248314 6860d71 c0222a9 17185c7 2ffb3d9 e05077f 2bcb56e d56e4d5 fbe6c16 4f6b708 a6d7f0c ecb302f 39d2f84 063a5a1 e488fa6 9cf0952 df1aa2d 242bc3c 7164c88 7a2e959 2af0efe 0749f2c 9d44cb7 49a32bd 5f40cfd a963305 f67093c 735e05a 29db87d 1ffd38f e982839 ddb08d4 c3c6165 34c9d7d cd349b7 ffe4763 9981c8c 693fa76 3d61d01 09e88af 26674d1 af50123 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 |
Started at: 10:42:59 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.2707270306348801 - MAE: 0.4014548746281838 Validation loss : 0.18827014460283167 - MAE: 0.3350395914193632 Epoch: 1 Training loss: 0.18397963613271714 - MAE: 0.3274942520732981 Validation loss : 0.17896055035731373 - MAE: 0.32625804300542655 Epoch: 2 Training loss: 0.17076618894934653 - MAE: 0.3156013590381467 Validation loss : 0.17423390377970302 - MAE: 0.32323377271249604 Epoch: 3 Training loss: 0.1643083082139492 - MAE: 0.3095641840936019 Validation loss : 0.16728906508754282 - MAE: 0.3149824083534162 Epoch: 4 Training loss: 0.15957146167755126 - MAE: 0.3037565492992936 Validation loss : 0.16289721429347992 - MAE: 0.3089864069106985 Epoch: 5 Training loss: 0.15854436449706555 - MAE: 0.301500893432691 Validation loss : 0.16129758033682318 - MAE: 0.308202317708313 Epoch: 6 Training loss: 0.15570181787014006 - MAE: 0.29918709351469924 Validation loss : 0.16344413292758606 - MAE: 0.30938825983369334 Epoch: 7 Training loss: 0.15292704798281193 - MAE: 0.2961117975066867 Validation loss : 0.16081692103077383 - MAE: 0.3033625257590202 Epoch: 8 Training loss: 0.15216302141547203 - MAE: 0.2959306926556599 Validation loss : 0.15965510378865636 - MAE: 0.30640949969727455 Epoch: 9 Training loss: 0.1456899941712618 - MAE: 0.2888247514317626 Validation loss : 0.15742756864603827 - MAE: 0.3014585494849406 Epoch: 10 Training loss: 0.1467710939049721 - MAE: 0.2907957975079187 Validation loss : 0.15912020206451416 - MAE: 0.30151619716495687 Epoch: 11 Training loss: 0.14659826673567294 - MAE: 0.2893981534314371 Validation loss : 0.1602539621731814 - MAE: 0.3022486893374817 Epoch: 12 Training loss: 0.14781650044023992 - MAE: 0.29275159014985785 Validation loss : 0.15743818177896388 - MAE: 0.3011467784998341 Epoch: 13 Training loss: 0.14709322392940521 - MAE: 0.2905235164202929 Validation loss : 0.15774143706349766 - MAE: 0.30035432342912505 Epoch: 14 Training loss: 0.14297918483614921 - MAE: 0.2870641656789985 Validation loss : 0.16193263846285202 - MAE: 0.3045518551046045 Epoch: 15 Training loss: 0.14410636112093925 - MAE: 0.2889387564652954 Validation loss : 0.15722107492825566 - MAE: 0.299146052737687 Epoch: 16 Training loss: 0.14345928631722926 - MAE: 0.28585450120342 Validation loss : 0.15835655086180744 - MAE: 0.2999799426627903 Epoch: 17 |