File size: 4,107 Bytes
0f4ddf6 94c1893 66c6131 3b94107 fbf603f 56fe2d3 1e73aec 20f5ed8 bc9be57 86c79cd 607bfff 1783dd2 4ee6ab4 d075457 9c99b98 4353f90 8845df1 2833e42 9acac77 fdc5a9f 63ed4ae e9d9b2d 0f5241d 4a87de6 5fe2307 2e43549 8b0f13e 6a992f2 670fd0d 4c2f141 dc71a12 e123fbc 325fe9d fd3c27c 4c9695c 20ca70e 8d4c8f6 c19c625 9b84df8 52617dc cf34feb 76518ab 69c74bc 78f22ab 54967e1 8b8f5fe 20d4375 7979e9f 4a0e93d 38437fd 33f3daf 950da3c 705a3dc 03b6e28 7b0bbd4 334f1aa d4299e0 a15807b fb9db54 bd463e7 4e15208 48e1193 e868138 f3e47f4 394f243 c369ae7 440aa5c ed96971 2d9c75e fe11629 abd7d78 2595a0e 21a9c8f caf392f 62192e6 671a46a dfc3aed b1437fe 73719d1 7b64bb1 6b42eb7 3a93cd6 a90ce90 7a098dc |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 |
Started at: 15:20:57 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.3436830922961235 - MAE: 0.4572523557050669 Validation loss : 0.26256299018859863 - MAE: 0.40728324728553433 Epoch: 1 Training loss: 0.21011242344975473 - MAE: 0.35190288965025124 Validation loss : 0.18477652966976166 - MAE: 0.3296733541388515 Epoch: 2 Training loss: 0.19629175141453742 - MAE: 0.33688825701191527 Validation loss : 0.18152975397450583 - MAE: 0.32990742521562766 Epoch: 3 Training loss: 0.1831872671842575 - MAE: 0.3247871467745211 Validation loss : 0.17427935344832285 - MAE: 0.3213832702684575 Epoch: 4 Training loss: 0.1773839347064495 - MAE: 0.31992620730258825 Validation loss : 0.17429945511477335 - MAE: 0.3230458382996596 Epoch: 5 Training loss: 0.17231015339493752 - MAE: 0.31171345188658794 Validation loss : 0.17113730950014933 - MAE: 0.31939983255844806 Epoch: 6 Training loss: 0.16903775408864022 - MAE: 0.3102792424124377 Validation loss : 0.16909515857696533 - MAE: 0.3172391366407421 Epoch: 7 Training loss: 0.16602707877755166 - MAE: 0.3075343290458685 Validation loss : 0.16411419212818146 - MAE: 0.31039227117874574 Epoch: 8 Training loss: 0.1662418097257614 - MAE: 0.3082071055792662 Validation loss : 0.16296945938042232 - MAE: 0.3092928071837739 Epoch: 9 Training loss: 0.16653021723031997 - MAE: 0.30672722178327916 Validation loss : 0.16100754695279257 - MAE: 0.30576533515417514 Epoch: 10 Training loss: 0.16153860688209534 - MAE: 0.30234728984229753 Validation loss : 0.16281781026295253 - MAE: 0.3046462940190389 Epoch: 11 Training loss: 0.16209525130689145 - MAE: 0.3039236849786875 Validation loss : 0.1614038199186325 - MAE: 0.30358997656721776 Epoch: 12 Training loss: 0.15954031608998775 - MAE: 0.3032080887111098 Validation loss : 0.16037237431321824 - MAE: 0.3024962801353889 Epoch: 13 Training loss: 0.1593211904168129 - MAE: 0.30014842639769856 Validation loss : 0.1610377154179982 - MAE: 0.30269789015625514 Epoch: 14 Training loss: 0.15742682591080664 - MAE: 0.2990433643869133 Validation loss : 0.16269906503813608 - MAE: 0.3034925004658223 Epoch: 15 Training loss: 0.15586433932185173 - MAE: 0.29719735232655653 Validation loss : 0.1581807051386152 - MAE: 0.3014812709966527 Epoch: 16 Training loss: 0.15324377864599228 - MAE: 0.29518426166093065 Validation loss : 0.15657826832362584 - MAE: 0.300699633359451 Epoch: 17 Training loss: 0.15139634534716606 - MAE: 0.2926187599905815 Validation loss : 0.15587361582687922 - MAE: 0.2989909241765162 Epoch: 18 Training loss: 0.150983314961195 - MAE: 0.2925825479179498 Validation loss : 0.15516561269760132 - MAE: 0.29856372533592385 Epoch: 19 Training loss: 0.1493927862495184 - MAE: 0.2910948114445842 Validation loss : 0.15466026110308512 - MAE: 0.2976538621938452 Epoch: 20 Training loss: 0.14843332655727864 - MAE: 0.2885189939815885 Validation loss : 0.15413911640644073 - MAE: 0.29834344421420217 Epoch: 21 Training loss: 0.1489902339875698 - MAE: 0.2893552038446195 Validation loss : 0.15573792159557343 - MAE: 0.29749915735213595 Epoch: 22 Training loss: 0.14985780157148837 - MAE: 0.29154637065109446 Validation loss : 0.15332231564181192 - MAE: 0.29697695339866076 Epoch: 23 Training loss: 0.14575221315026282 - MAE: 0.2886255141256018 Validation loss : 0.15240811024393355 - MAE: 0.29550651769807773 Epoch: 24 Training loss: 0.14468684382736682 - MAE: 0.2861701193601813 Validation loss : 0.15227091951029642 - MAE: 0.2969623801494556 Epoch: 25 Training loss: 0.14725722186267376 - MAE: 0.28829223441788526 Validation loss : 0.15270294461931502 - MAE: 0.29577565945239215 Epoch: 26 Training loss: 0.14587189629673958 - MAE: 0.2862314013226889 Validation loss : 0.1532691674573081 - MAE: 0.29749263015344235 Stopped after 26 epochs. |