File size: 10,622 Bytes
a9614e2 1c266c2 573d184 138c452 085fc0e a8e4a32 60030c6 43aca8f bced6f4 3c0ba3e 05cb146 2755fdc 6d43361 9ba6b1c bc4b6fa fd66402 8ac5557 0a5b886 7582378 7943a4b 1d207e4 fc0b9fb 778b2ee d8f234d b8f30b4 07d6b01 b1e9821 5d63be6 c240e23 7fd6a30 753d906 7988c17 f9cc4b4 2563dd5 3b32d9d 78c220f 13dd3c9 3645ffc a67ad3c 1791d1d 417295b 9d01042 4692ce2 ee702fe 9ad9b31 8a55390 af3dc92 255d975 b5c59b2 6291ba6 910096e f1cb193 5ecc42f f52fd05 4aa4325 5b2f689 0ca188f 90fba87 3b6adf1 d7b4484 b3d9166 822020b 5f1eb05 5a03441 36bc79e ff030dd 927b3c1 64f46f0 749183e 2c6fb37 bff4a2d 46697fc 0413f7a 5ee6dd5 419fd28 eb700da 64c500b d0056fd 66982b8 1d41cd7 1c356ba d930327 fcc9b1f 1399e97 04ba03c ed8b3f4 025040b 13b93d7 d26ef5a a37fc5f cf6faf1 86a6afb 40f3e9f 7cbbdc1 f5369b2 63b151c eaadc50 6354a95 0d8ed18 8182faf f9e751c cb1f440 8a3fa68 21062cf 7ce2a2c 1f989a1 3d97e81 23665ab 57ada11 4a041ea 91a97e6 f4e0960 42a05eb 214e4e3 a4c8ea3 1d48055 55af590 864f0d5 d12921d b7fd40a e4b2e16 35063ea c287647 9cc161b 16df9c1 8c239b2 e1ffb75 f067b6b 5707b55 06dbd2d ba89237 748bfcb ce6bb9d a0c316e 0e41bee a7c336c 284862e e6bfda7 ff1b253 7f85e3c 39360cd bf774fd 751781f 13f6f78 a14234f 2e1113b 26dbf57 48ecfcf c37f189 5e1b947 76ec1ac bc66fa8 71741a1 6fe9bdc cba9e16 e6484a9 e9f9dc2 e2560d5 94d08de ab2ab8b 306c777 b0daf82 7161dc2 ec8b412 6e7f355 1dbcf89 74c5bce 8f28059 cd5e1ba 50d54b4 e3854b6 47f5e22 0e66be8 0596de2 67a7610 eacd439 610aa7d 76212f8 5a43060 94586e7 6fae1c1 8ce35ef de84ad0 71f0a6e 7988ea4 2ce4e82 0b0e4ff e51065c c85d65b be9081c d27e527 4fe4ad7 edff910 8f89e59 f773786 0c18394 e5713c6 4257128 587cfee 2855675 c951d8f 26d82b2 668d135 cf661ef 5d6ba54 e07588e 0bf1606 7ba5a4f 5ccf23e 04c8203 3ce5140 0142e2d aa8c5f1 14e679e 5e114da 3fa5453 2d659e3 d996cf0 56ae8c1 0ff9e6e cef854e db07dce b013ee1 af9d814 c9d4c79 270f1db 0a8d8f5 757246c b4908df |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 |
Started at: 21:34:05 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.32175946950912476 - MAE: 0.4437903503551826 Validation loss : 0.19621961481041378 - MAE: 0.3388283822410594 Epoch: 1 Training loss: 0.21072698533535003 - MAE: 0.35116530289867337 Validation loss : 0.1876767095592287 - MAE: 0.33628140968919695 Epoch: 2 Training loss: 0.1893942326307297 - MAE: 0.3312522516218666 Validation loss : 0.1748287214173211 - MAE: 0.3213797447248068 Epoch: 3 Training loss: 0.17969159841537474 - MAE: 0.3215382240664255 Validation loss : 0.1697687026527193 - MAE: 0.31446870224950263 Epoch: 4 Training loss: 0.17220519423484804 - MAE: 0.3129526796594085 Validation loss : 0.16811848680178323 - MAE: 0.3117262145282266 Epoch: 5 Training loss: 0.17296144247055054 - MAE: 0.31494873624825814 Validation loss : 0.16849000917540657 - MAE: 0.31104514179599985 Epoch: 6 Training loss: 0.16641035348176955 - MAE: 0.30768475494362546 Validation loss : 0.16421516074074638 - MAE: 0.30715220958263423 Epoch: 7 Training loss: 0.16057054400444032 - MAE: 0.30329494898002457 Validation loss : 0.161027698053254 - MAE: 0.304297376785587 Epoch: 8 Training loss: 0.1566170272231102 - MAE: 0.29957014870206655 Validation loss : 0.15982638630602095 - MAE: 0.30397747682822174 Epoch: 9 Training loss: 0.15673983812332154 - MAE: 0.3000540458404174 Validation loss : 0.159184659520785 - MAE: 0.3035235378542429 Epoch: 10 Training loss: 0.15630604147911073 - MAE: 0.29690365842430627 Validation loss : 0.15921704471111298 - MAE: 0.30240467396157955 Epoch: 11 Training loss: 0.15595020622015 - MAE: 0.29754135005638765 Validation loss : 0.15686986181471083 - MAE: 0.30119996351152656 Epoch: 12 Training loss: 0.15274528950452804 - MAE: 0.2944161972508913 Validation loss : 0.15652166141404045 - MAE: 0.3012713923501961 Epoch: 13 Training loss: 0.15205995708703995 - MAE: 0.29424324063629004 Validation loss : 0.1575678288936615 - MAE: 0.3004794443503115 Epoch: 14 Training loss: 0.15087180227041244 - MAE: 0.292504579327589 Validation loss : 0.15574459234873453 - MAE: 0.29932356111019714 Epoch: 15 Training loss: 0.1518820345401764 - MAE: 0.2940904971897683 Validation loss : 0.15581322544150883 - MAE: 0.30048684662514935 Epoch: 16 Training loss: 0.14916340589523316 - MAE: 0.2900369708605816 Validation loss : 0.15502946575482687 - MAE: 0.2986623058669649 Epoch: 17 Training loss: 0.14969733864068985 - MAE: 0.2916869417468108 Validation loss : 0.15631223718325296 - MAE: 0.2994015598567933 Epoch: 18 Training loss: 0.14621972769498826 - MAE: 0.2875086269286061 Validation loss : 0.1557358337773217 - MAE: 0.29931970436403404 Epoch: 19 Training loss: 0.14798570185899734 - MAE: 0.29024787778757843 Validation loss : 0.15424930387073094 - MAE: 0.2984811820958494 Epoch: 20 Training loss: 0.14769238144159316 - MAE: 0.2896109423923894 Validation loss : 0.1536422868569692 - MAE: 0.2987445844262027 Epoch: 21 Training loss: 0.14361368536949157 - MAE: 0.2842206176667335 Validation loss : 0.1533755792511834 - MAE: 0.29642898867488277 Epoch: 22 Training loss: 0.14456430345773696 - MAE: 0.2860957867005398 Validation loss : 0.1542035871081882 - MAE: 0.2971775973100257 Epoch: 23 Training loss: 0.1426533755660057 - MAE: 0.28512021628758083 Validation loss : 0.15245803362793392 - MAE: 0.2961228783995986 Epoch: 24 Training loss: 0.14542177826166153 - MAE: 0.2855342243309425 Validation loss : 0.15294104317824045 - MAE: 0.29643245601047447 Epoch: 25 Training loss: 0.14522234290838243 - MAE: 0.2851345876886818 Validation loss : 0.15342944694889915 - MAE: 0.2958033843008907 Epoch: 26 Training loss: 0.14386597275733948 - MAE: 0.2838243646140143 Validation loss : 0.1515038808186849 - MAE: 0.2936464496284943 Epoch: 27 Training loss: 0.1415349954366684 - MAE: 0.2810586437104536 Validation loss : 0.1507236527072059 - MAE: 0.29409404478735646 Epoch: 28 Training loss: 0.14115683376789093 - MAE: 0.2815221038197439 Validation loss : 0.15061336921321022 - MAE: 0.29280129784932213 Epoch: 29 Training loss: 0.14186541587114335 - MAE: 0.28239653007887255 Validation loss : 0.15151139597098032 - MAE: 0.2953867482779256 Epoch: 30 Training loss: 0.14276256740093232 - MAE: 0.283208699006944 Validation loss : 0.1504600097735723 - MAE: 0.29276670315056585 Epoch: 31 Training loss: 0.14241950929164887 - MAE: 0.28277882078383293 Validation loss : 0.1505332812666893 - MAE: 0.29293145394981956 Epoch: 32 Training loss: 0.14204519629478454 - MAE: 0.28192363607736504 Validation loss : 0.15044088496102226 - MAE: 0.2929076596067834 Epoch: 33 Training loss: 0.13920597046613692 - MAE: 0.2793298976614221 Validation loss : 0.15006180769867367 - MAE: 0.29292432343849906 Epoch: 34 Training loss: 0.1410813584923744 - MAE: 0.2815379053911258 Validation loss : 0.14930015967951882 - MAE: 0.2933547268525188 Epoch: 35 Training loss: 0.14150760889053346 - MAE: 0.2832031330097777 Validation loss : 0.15295125875208113 - MAE: 0.2936935655701593 Epoch: 36 Training loss: 0.1405733221769333 - MAE: 0.2821643593441243 Validation loss : 0.14867112785577774 - MAE: 0.2921335742814868 Epoch: 37 Training loss: 0.14206359952688216 - MAE: 0.28261241361934286 Validation loss : 0.14988169405195448 - MAE: 0.2927604772074144 Epoch: 38 Training loss: 0.14064243495464324 - MAE: 0.28185439579063915 Validation loss : 0.1515509072277281 - MAE: 0.29319069832718325 Epoch: 39 Training loss: 0.14297345608472825 - MAE: 0.282847817609895 Validation loss : 0.15042786465750801 - MAE: 0.29282722784595605 Epoch: 40 Training loss: 0.13990240722894667 - MAE: 0.27964537481988466 Validation loss : 0.14972211172183356 - MAE: 0.29260355981671565 Epoch: 41 Training loss: 0.13921964198350906 - MAE: 0.28078253547199195 Validation loss : 0.14965214083592096 - MAE: 0.2924700199853063 Epoch: 42 Training loss: 0.1393338233232498 - MAE: 0.28024463401390987 Validation loss : 0.1507701873779297 - MAE: 0.29355734742186085 Epoch: 43 Training loss: 0.1382669734954834 - MAE: 0.27992590956331337 Validation loss : 0.14967524343066746 - MAE: 0.29149855699855537 Epoch: 44 Training loss: 0.14029370576143266 - MAE: 0.28265328374425863 Validation loss : 0.1491319098406368 - MAE: 0.29219012845466735 Epoch: 45 Training loss: 0.14041882872581482 - MAE: 0.2809566443072132 Validation loss : 0.14969545520014232 - MAE: 0.2909081862153574 Epoch: 46 Training loss: 0.140560123026371 - MAE: 0.28191663739835304 Validation loss : 0.14977325167920855 - MAE: 0.291604275381065 Epoch: 47 Training loss: 0.1414617270231247 - MAE: 0.28282842532543745 Validation loss : 0.1506698860062493 - MAE: 0.2930552446484735 Epoch: 48 Training loss: 0.1373920688033104 - MAE: 0.27917131881687096 Validation loss : 0.15224417712953356 - MAE: 0.2934837599479946 Epoch: 49 Training loss: 0.14192317247390748 - MAE: 0.2842585649178634 Validation loss : 0.14885787086354363 - MAE: 0.29192342005218896 Epoch: 50 Training loss: 0.14062062561511993 - MAE: 0.2808979887022064 Validation loss : 0.15054219878382152 - MAE: 0.2924492759311905 Epoch: 51 Training loss: 0.13990722686052323 - MAE: 0.2817114196110173 Validation loss : 0.1506080221798685 - MAE: 0.29247582518131027 Epoch: 52 Training loss: 0.1394276064634323 - MAE: 0.2798831110400087 Validation loss : 0.1490704102648629 - MAE: 0.291844280782309 Epoch: 53 Training loss: 0.14104595750570298 - MAE: 0.28180114573645876 Validation loss : 0.14941397143734825 - MAE: 0.2933880270519655 Epoch: 54 Training loss: 0.139021515250206 - MAE: 0.2793780218867486 Validation loss : 0.1490028988983896 - MAE: 0.29192548589513484 Epoch: 55 Training loss: 0.1409096112847328 - MAE: 0.2816018781967446 Validation loss : 0.1494157827562756 - MAE: 0.2919783378546487 Epoch: 56 Training loss: 0.14150582283735275 - MAE: 0.28397171057514964 Validation loss : 0.1507257032725546 - MAE: 0.29278463527362475 Epoch: 57 Training loss: 0.1412457060813904 - MAE: 0.28194929681996234 Validation loss : 0.1500752924217118 - MAE: 0.291890501089955 Epoch: 58 Training loss: 0.14060731261968612 - MAE: 0.2826944446609872 Validation loss : 0.15134404020176995 - MAE: 0.2921265934844939 Epoch: 59 Training loss: 0.13857035160064698 - MAE: 0.27917573503863885 Validation loss : 0.1502113996280564 - MAE: 0.2916726146711328 Epoch: 60 Training loss: 0.14019066840410233 - MAE: 0.2816340620891049 Validation loss : 0.14821443955103555 - MAE: 0.29117830198396194 Epoch: 61 Training loss: 0.1397022610902786 - MAE: 0.28089572600032964 Validation loss : 0.1513935931854778 - MAE: 0.2917060470497501 Epoch: 62 Training loss: 0.14070956617593766 - MAE: 0.28241979367979564 Validation loss : 0.1501304821835624 - MAE: 0.2912750997306758 Epoch: 63 Training loss: 0.1420893755555153 - MAE: 0.28372013668513324 Validation loss : 0.15038747092088064 - MAE: 0.2920179987433622 Epoch: 64 Training loss: 0.1394691723585129 - MAE: 0.2797968922342149 Validation loss : 0.1498954254719946 - MAE: 0.2930247953251004 Epoch: 65 Training loss: 0.14013640880584716 - MAE: 0.28057332071413116 Validation loss : 0.14924369752407074 - MAE: 0.29117805094765137 Epoch: 66 Training loss: 0.13910346567630769 - MAE: 0.27976148521760813 Validation loss : 0.15365933709674412 - MAE: 0.2938181280395696 Epoch: 67 Training loss: 0.14027394950389863 - MAE: 0.2824978199508478 Validation loss : 0.15143409785297182 - MAE: 0.2915787634959989 Epoch: 68 Training loss: 0.13879832834005357 - MAE: 0.2803663281034843 Validation loss : 0.14806738330258262 - MAE: 0.2899341093266117 Epoch: 69 Training loss: 0.14169045120477677 - MAE: 0.2840059654564348 Validation loss : 0.14932368530167472 - MAE: 0.2908179996636204 Epoch: 70 Training loss: 0.14286602079868316 - MAE: 0.28393754014878414 Validation loss : 0.15445275604724884 - MAE: 0.29465461309717 Epoch: 71 Training loss: 0.1397111463546753 - MAE: 0.2815673786321848 Validation loss : 0.14909260223309198 - MAE: 0.29124006221654514 Epoch: 72 Training loss: 0.13914978861808777 - MAE: 0.27919056115617275 Validation loss : 0.1496173938115438 - MAE: 0.29145609362558494 Epoch: 73 Training loss: 0.13840563476085663 - MAE: 0.2791765280391529 Validation loss : 0.15138710704114702 - MAE: 0.2922445451728484 Epoch: 74 Training loss: 0.13956822723150253 - MAE: 0.2806497885355469 Validation loss : 0.14996923257907233 - MAE: 0.2913337573766691 Epoch: 75 Training loss: 0.13771951258182524 - MAE: 0.2787076577823987 |