File size: 6,907 Bytes
a9476ac 2f56b4a 2888120 3184f37 8ae7bc4 e541756 feac5b7 3ddbab6 f146e31 84c4479 adda1e0 b0b276b d64d87a 3174ab2 921fb61 a26e9c5 7512ab2 97f1067 d55f347 a8ec985 c4cc5a8 42fb0a1 208b67e 65c19a7 da2e07b b42b958 5505eb5 008f825 d7f5617 901daef a093e18 065aaa2 ccad6e6 5bfc86d 6f1770d c8e56ff 9c08539 0477fe3 5ad6ace 66aef13 0d8fd65 5cd05c3 cd71def 26b9a28 f2cd962 f6298b5 6aba17d 21c851f 96ec8b8 5df4046 3e16a65 9289b26 3fecdfd 9067e1b 250b7a7 fb1935f 01ea17d 04dcb73 fa24b9f 72c7f31 8e4bf7d b0de0a7 b0194cc daae036 7590205 5952910 2c72d62 d4d9ee8 1b27b06 2959328 bb1e171 e10d4b0 0767102 f13d34f 0794506 52ca4e9 6cc1111 9555a49 a1c1611 ece3171 cf4b686 425f182 091f24a 58df72a 71bd9ce 517b39f 9831719 7126ff0 e32397c a974ca1 a8698d1 24b8e96 a1ede8c 47b914a 65aa7a8 2959bcf a7435ed ea88322 58f6c77 d210501 b5346f7 969e355 51ab487 3492889 694ca3e 0ab8169 3741901 d8c9e88 b7f34e1 3021af6 2109733 b67ad55 1f9643f 6d318cc c6b762b 9188c77 0edd6e1 e29056e b70a629 256e9f5 e4bf36e fbeaf4f 4844ce7 e5f5fc5 a81b140 5b9d94e 0ac473c cf173eb fd1123c b36d6e2 4077e9f 38e53d1 603e643 90f8921 9509bb4 dc65ca0 265dfee a105a8f 29c4134 4dd2cb6 db649d1 254cc32 b898f80 738f1ee a22db52 095b0b1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 |
Started at: 12:12:18 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 1.1006720215082169 - MAE: 0.9385178221693342 Validation loss : 0.3364674378843868 - MAE: 0.4550059916177428 Epoch: 1 Training loss: 0.2378660583496094 - MAE: 0.37276055966635346 Validation loss : 0.1963765437112135 - MAE: 0.34272739887256237 Epoch: 2 Training loss: 0.19169337004423143 - MAE: 0.3349637829761448 Validation loss : 0.1931756366701687 - MAE: 0.34055783097602843 Epoch: 3 Training loss: 0.18910380512475966 - MAE: 0.3318428617734502 Validation loss : 0.18860271222451153 - MAE: 0.3358887044598495 Epoch: 4 Training loss: 0.186059138327837 - MAE: 0.32876787579915384 Validation loss : 0.1856619251124999 - MAE: 0.33292138410324873 Epoch: 5 Training loss: 0.18313051164150237 - MAE: 0.32580828382716187 Validation loss : 0.1833767566610785 - MAE: 0.3306058463127525 Epoch: 6 Training loss: 0.18097147583961487 - MAE: 0.32398510773591493 Validation loss : 0.1812712576459436 - MAE: 0.3284241419338835 Epoch: 7 Training loss: 0.17751911029219627 - MAE: 0.32037782771308876 Validation loss : 0.17958090322859147 - MAE: 0.3267873033045649 Epoch: 8 Training loss: 0.1786111931502819 - MAE: 0.321583163490193 Validation loss : 0.17803318272618687 - MAE: 0.32520436183204793 Epoch: 9 Training loss: 0.17398411139845849 - MAE: 0.31673606421156436 Validation loss : 0.17673131735885844 - MAE: 0.3238730944125878 Epoch: 10 Training loss: 0.17270002499222756 - MAE: 0.31745671014834814 Validation loss : 0.17539521964157329 - MAE: 0.3224941170700243 Epoch: 11 Training loss: 0.17052328288555146 - MAE: 0.3148727480077667 Validation loss : 0.174127253539422 - MAE: 0.3211262794601334 Epoch: 12 Training loss: 0.16809561610221863 - MAE: 0.311770730424323 Validation loss : 0.1734507837716271 - MAE: 0.3205848137252918 Epoch: 13 Training loss: 0.16760414585471153 - MAE: 0.3136055973164791 Validation loss : 0.17252265530474045 - MAE: 0.3196495827839378 Epoch: 14 Training loss: 0.1682390233874321 - MAE: 0.3112908435346189 Validation loss : 0.1711487393168842 - MAE: 0.31795443358982844 Epoch: 15 Training loss: 0.16781458392739296 - MAE: 0.3113590658699254 Validation loss : 0.1707548902315252 - MAE: 0.3177303930498194 Epoch: 16 Training loss: 0.16693725898861886 - MAE: 0.31030240300129464 Validation loss : 0.16986496132962844 - MAE: 0.3166979421895762 Epoch: 17 Training loss: 0.16504903733730317 - MAE: 0.31016979911031045 Validation loss : 0.16925524350474863 - MAE: 0.31608280831472446 Epoch: 18 Training loss: 0.16303982958197594 - MAE: 0.30655030591901256 Validation loss : 0.1685786873978727 - MAE: 0.3153196465443914 Epoch: 19 Training loss: 0.1638409486413002 - MAE: 0.3080217884644864 Validation loss : 0.16801373134641087 - MAE: 0.314735410389949 Epoch: 20 Training loss: 0.1614493829011917 - MAE: 0.30555361713211904 Validation loss : 0.1675474293091718 - MAE: 0.3142234109718509 Epoch: 21 Training loss: 0.16097557842731475 - MAE: 0.30540769164567677 Validation loss : 0.1668663542060291 - MAE: 0.31328988497467153 Epoch: 22 Training loss: 0.15972515404224397 - MAE: 0.3035523077998348 Validation loss : 0.16639567999278798 - MAE: 0.31270315476408916 Epoch: 23 Training loss: 0.15968465387821199 - MAE: 0.3035132420493961 Validation loss : 0.16612719788270838 - MAE: 0.3124643849701637 Epoch: 24 Training loss: 0.16047569021582603 - MAE: 0.3044606072582237 Validation loss : 0.16531289544175654 - MAE: 0.3113139200118973 Epoch: 25 Training loss: 0.1594112578034401 - MAE: 0.30380153537656146 Validation loss : 0.16510769549538107 - MAE: 0.31113827420553686 Epoch: 26 Training loss: 0.15761824071407318 - MAE: 0.30175219485110094 Validation loss : 0.16493081169969895 - MAE: 0.3109766566528519 Epoch: 27 Training loss: 0.15642025277018548 - MAE: 0.2996838668675141 Validation loss : 0.16460545141907298 - MAE: 0.3105549194584687 Epoch: 28 Training loss: 0.1564825715124607 - MAE: 0.29999498344721554 Validation loss : 0.16410903238198338 - MAE: 0.3098771926366311 Epoch: 29 Training loss: 0.15597251012921334 - MAE: 0.3000410448533109 Validation loss : 0.16374634469256683 - MAE: 0.3094041646235353 Epoch: 30 Training loss: 0.15675333008170128 - MAE: 0.3009904709763715 Validation loss : 0.16385302824132583 - MAE: 0.30973321130028236 Epoch: 31 Training loss: 0.15694587096571921 - MAE: 0.2996392168091802 Validation loss : 0.16349118728848064 - MAE: 0.3092097406182407 Epoch: 32 Training loss: 0.15658894553780556 - MAE: 0.2999551267813354 Validation loss : 0.16314544002799428 - MAE: 0.308750879464887 Epoch: 33 Training loss: 0.15591454565525054 - MAE: 0.3000454994119826 Validation loss : 0.16286137261811426 - MAE: 0.30841066500734077 Epoch: 34 Training loss: 0.15509203642606736 - MAE: 0.29930140182204795 Validation loss : 0.16284336149692535 - MAE: 0.3085285183124655 Epoch: 35 Training loss: 0.15405697703361512 - MAE: 0.2988630300365712 Validation loss : 0.16246377238455942 - MAE: 0.3079503563975846 Epoch: 36 Training loss: 0.15352507963776588 - MAE: 0.29748678107055193 Validation loss : 0.16194848672432058 - MAE: 0.3071383866301113 Epoch: 37 Training loss: 0.15278249979019165 - MAE: 0.2962795231384297 Validation loss : 0.16206031862427206 - MAE: 0.30743531882524455 Epoch: 38 Training loss: 0.15470703780651093 - MAE: 0.2978209194206603 Validation loss : 0.16202253015602336 - MAE: 0.30753639777984476 Epoch: 39 Training loss: 0.1523985606431961 - MAE: 0.2964090176553577 Validation loss : 0.16172013650922215 - MAE: 0.30709932655101335 Epoch: 40 Training loss: 0.1530853621661663 - MAE: 0.297757612122841 Validation loss : 0.1613460998324787 - MAE: 0.30660857549224224 Epoch: 41 Training loss: 0.15282656386494636 - MAE: 0.297049206683197 Validation loss : 0.1608997815672089 - MAE: 0.30598334249346854 Epoch: 42 Training loss: 0.15108009725809096 - MAE: 0.29496736148642777 Validation loss : 0.16095499475212657 - MAE: 0.3062596970514697 Epoch: 43 Training loss: 0.15141472831368447 - MAE: 0.29662705431765973 Validation loss : 0.16070682117167642 - MAE: 0.30597615449664417 Epoch: 44 Training loss: 0.15191791325807572 - MAE: 0.2950182572471767 Validation loss : 0.1603742417167215 - MAE: 0.30558954112078246 Epoch: 45 Training loss: 0.15128972373902796 - MAE: 0.29699661967263946 Validation loss : 0.159849391264074 - MAE: 0.30474853877294245 Epoch: 46 Training loss: 0.15125623732805252 - MAE: 0.2951512536815187 Validation loss : 0.16034438154276678 - MAE: 0.3057817503470798 Epoch: 47 Training loss: 0.1518179516494274 - MAE: 0.2959772343677161 Validation loss : 0.16012860746944652 - MAE: 0.3054851252496958 |