File size: 2,665 Bytes
0f4ddf6 94c1893 66c6131 3b94107 fbf603f 56fe2d3 1e73aec 20f5ed8 bc9be57 86c79cd 607bfff 1783dd2 4ee6ab4 d075457 9c99b98 4353f90 8845df1 2833e42 9acac77 fdc5a9f 63ed4ae e9d9b2d 0f5241d 4a87de6 5fe2307 2e43549 8b0f13e 6a992f2 670fd0d 4c2f141 dc71a12 e123fbc 325fe9d fd3c27c 4c9695c 20ca70e 8d4c8f6 c19c625 9b84df8 52617dc cf34feb 76518ab 69c74bc 78f22ab 54967e1 8b8f5fe 20d4375 7979e9f 4a0e93d 38437fd 33f3daf 950da3c |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 |
Started at: 15:20:57 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.3436830922961235 - MAE: 0.4572523557050669 Validation loss : 0.26256299018859863 - MAE: 0.40728324728553433 Epoch: 1 Training loss: 0.21011242344975473 - MAE: 0.35190288965025124 Validation loss : 0.18477652966976166 - MAE: 0.3296733541388515 Epoch: 2 Training loss: 0.19629175141453742 - MAE: 0.33688825701191527 Validation loss : 0.18152975397450583 - MAE: 0.32990742521562766 Epoch: 3 Training loss: 0.1831872671842575 - MAE: 0.3247871467745211 Validation loss : 0.17427935344832285 - MAE: 0.3213832702684575 Epoch: 4 Training loss: 0.1773839347064495 - MAE: 0.31992620730258825 Validation loss : 0.17429945511477335 - MAE: 0.3230458382996596 Epoch: 5 Training loss: 0.17231015339493752 - MAE: 0.31171345188658794 Validation loss : 0.17113730950014933 - MAE: 0.31939983255844806 Epoch: 6 Training loss: 0.16903775408864022 - MAE: 0.3102792424124377 Validation loss : 0.16909515857696533 - MAE: 0.3172391366407421 Epoch: 7 Training loss: 0.16602707877755166 - MAE: 0.3075343290458685 Validation loss : 0.16411419212818146 - MAE: 0.31039227117874574 Epoch: 8 Training loss: 0.1662418097257614 - MAE: 0.3082071055792662 Validation loss : 0.16296945938042232 - MAE: 0.3092928071837739 Epoch: 9 Training loss: 0.16653021723031997 - MAE: 0.30672722178327916 Validation loss : 0.16100754695279257 - MAE: 0.30576533515417514 Epoch: 10 Training loss: 0.16153860688209534 - MAE: 0.30234728984229753 Validation loss : 0.16281781026295253 - MAE: 0.3046462940190389 Epoch: 11 Training loss: 0.16209525130689145 - MAE: 0.3039236849786875 Validation loss : 0.1614038199186325 - MAE: 0.30358997656721776 Epoch: 12 Training loss: 0.15954031608998775 - MAE: 0.3032080887111098 Validation loss : 0.16037237431321824 - MAE: 0.3024962801353889 Epoch: 13 Training loss: 0.1593211904168129 - MAE: 0.30014842639769856 Validation loss : 0.1610377154179982 - MAE: 0.30269789015625514 Epoch: 14 Training loss: 0.15742682591080664 - MAE: 0.2990433643869133 Validation loss : 0.16269906503813608 - MAE: 0.3034925004658223 Epoch: 15 Training loss: 0.15586433932185173 - MAE: 0.29719735232655653 Validation loss : 0.1581807051386152 - MAE: 0.3014812709966527 Epoch: 16 Training loss: 0.15324377864599228 - MAE: 0.29518426166093065 |