Started at: 21:34:05 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.32175946950912476 - MAE: 0.4437903503551826 Validation loss : 0.19621961481041378 - MAE: 0.3388283822410594 Epoch: 1 Training loss: 0.21072698533535003 - MAE: 0.35116530289867337 Validation loss : 0.1876767095592287 - MAE: 0.33628140968919695 Epoch: 2 Training loss: 0.1893942326307297 - MAE: 0.3312522516218666 Validation loss : 0.1748287214173211 - MAE: 0.3213797447248068 Epoch: 3 Training loss: 0.17969159841537474 - MAE: 0.3215382240664255 Validation loss : 0.1697687026527193 - MAE: 0.31446870224950263 Epoch: 4 Training loss: 0.17220519423484804 - MAE: 0.3129526796594085 Validation loss : 0.16811848680178323 - MAE: 0.3117262145282266 Epoch: 5 Training loss: 0.17296144247055054 - MAE: 0.31494873624825814 Validation loss : 0.16849000917540657 - MAE: 0.31104514179599985 Epoch: 6 Training loss: 0.16641035348176955 - MAE: 0.30768475494362546 Validation loss : 0.16421516074074638 - MAE: 0.30715220958263423 Epoch: 7 Training loss: 0.16057054400444032 - MAE: 0.30329494898002457 Validation loss : 0.161027698053254 - MAE: 0.304297376785587 Epoch: 8 Training loss: 0.1566170272231102 - MAE: 0.29957014870206655 Validation loss : 0.15982638630602095 - MAE: 0.30397747682822174 Epoch: 9 Training loss: 0.15673983812332154 - MAE: 0.3000540458404174 Validation loss : 0.159184659520785 - MAE: 0.3035235378542429 Epoch: 10 Training loss: 0.15630604147911073 - MAE: 0.29690365842430627 Validation loss : 0.15921704471111298 - MAE: 0.30240467396157955 Epoch: 11 Training loss: 0.15595020622015 - MAE: 0.29754135005638765 Validation loss : 0.15686986181471083 - MAE: 0.30119996351152656 Epoch: 12 Training loss: 0.15274528950452804 - MAE: 0.2944161972508913 Validation loss : 0.15652166141404045 - MAE: 0.3012713923501961 Epoch: 13 Training loss: 0.15205995708703995 - MAE: 0.29424324063629004 Validation loss : 0.1575678288936615 - MAE: 0.3004794443503115 Epoch: 14 Training loss: 0.15087180227041244 - MAE: 0.292504579327589 Validation loss : 0.15574459234873453 - MAE: 0.29932356111019714 Epoch: 15 Training loss: 0.1518820345401764 - MAE: 0.2940904971897683 Validation loss : 0.15581322544150883 - MAE: 0.30048684662514935 Epoch: 16 Training loss: 0.14916340589523316 - MAE: 0.2900369708605816 Validation loss : 0.15502946575482687 - MAE: 0.2986623058669649 Epoch: 17 Training loss: 0.14969733864068985 - MAE: 0.2916869417468108 Validation loss : 0.15631223718325296 - MAE: 0.2994015598567933 Epoch: 18 Training loss: 0.14621972769498826 - MAE: 0.2875086269286061 Validation loss : 0.1557358337773217 - MAE: 0.29931970436403404 Epoch: 19 Training loss: 0.14798570185899734 - MAE: 0.29024787778757843 Validation loss : 0.15424930387073094 - MAE: 0.2984811820958494 Epoch: 20 Training loss: 0.14769238144159316 - MAE: 0.2896109423923894 Validation loss : 0.1536422868569692 - MAE: 0.2987445844262027 Epoch: 21 Training loss: 0.14361368536949157 - MAE: 0.2842206176667335 Validation loss : 0.1533755792511834 - MAE: 0.29642898867488277 Epoch: 22 Training loss: 0.14456430345773696 - MAE: 0.2860957867005398 Validation loss : 0.1542035871081882 - MAE: 0.2971775973100257 Epoch: 23 Training loss: 0.1426533755660057 - MAE: 0.28512021628758083 Validation loss : 0.15245803362793392 - MAE: 0.2961228783995986 Epoch: 24 Training loss: 0.14542177826166153 - MAE: 0.2855342243309425 Validation loss : 0.15294104317824045 - MAE: 0.29643245601047447 Epoch: 25 Training loss: 0.14522234290838243 - MAE: 0.2851345876886818 Validation loss : 0.15342944694889915 - MAE: 0.2958033843008907 Epoch: 26 Training loss: 0.14386597275733948 - MAE: 0.2838243646140143 Validation loss : 0.1515038808186849 - MAE: 0.2936464496284943 Epoch: 27 Training loss: 0.1415349954366684 - MAE: 0.2810586437104536 Validation loss : 0.1507236527072059 - MAE: 0.29409404478735646 Epoch: 28 Training loss: 0.14115683376789093 - MAE: 0.2815221038197439 Validation loss : 0.15061336921321022 - MAE: 0.29280129784932213 Epoch: 29 Training loss: 0.14186541587114335 - MAE: 0.28239653007887255 Validation loss : 0.15151139597098032 - MAE: 0.2953867482779256 Epoch: 30 Training loss: 0.14276256740093232 - MAE: 0.283208699006944 Validation loss : 0.1504600097735723 - MAE: 0.29276670315056585 Epoch: 31 Training loss: 0.14241950929164887 - MAE: 0.28277882078383293 Validation loss : 0.1505332812666893 - MAE: 0.29293145394981956 Epoch: 32 Training loss: 0.14204519629478454 - MAE: 0.28192363607736504 Validation loss : 0.15044088496102226 - MAE: 0.2929076596067834 Epoch: 33 Training loss: 0.13920597046613692 - MAE: 0.2793298976614221 Validation loss : 0.15006180769867367 - MAE: 0.29292432343849906 Epoch: 34 Training loss: 0.1410813584923744 - MAE: 0.2815379053911258 Validation loss : 0.14930015967951882 - MAE: 0.2933547268525188 Epoch: 35 Training loss: 0.14150760889053346 - MAE: 0.2832031330097777 Validation loss : 0.15295125875208113 - MAE: 0.2936935655701593 Epoch: 36 Training loss: 0.1405733221769333 - MAE: 0.2821643593441243 Validation loss : 0.14867112785577774 - MAE: 0.2921335742814868 Epoch: 37 Training loss: 0.14206359952688216 - MAE: 0.28261241361934286 Validation loss : 0.14988169405195448 - MAE: 0.2927604772074144 Epoch: 38 Training loss: 0.14064243495464324 - MAE: 0.28185439579063915 Validation loss : 0.1515509072277281 - MAE: 0.29319069832718325 Epoch: 39 Training loss: 0.14297345608472825 - MAE: 0.282847817609895 Validation loss : 0.15042786465750801 - MAE: 0.29282722784595605 Epoch: 40 Training loss: 0.13990240722894667 - MAE: 0.27964537481988466 Validation loss : 0.14972211172183356 - MAE: 0.29260355981671565 Epoch: 41 Training loss: 0.13921964198350906 - MAE: 0.28078253547199195 Validation loss : 0.14965214083592096 - MAE: 0.2924700199853063 Epoch: 42 Training loss: 0.1393338233232498 - MAE: 0.28024463401390987 Validation loss : 0.1507701873779297 - MAE: 0.29355734742186085 Epoch: 43 Training loss: 0.1382669734954834 - MAE: 0.27992590956331337 Validation loss : 0.14967524343066746 - MAE: 0.29149855699855537 Epoch: 44 Training loss: 0.14029370576143266 - MAE: 0.28265328374425863 Validation loss : 0.1491319098406368 - MAE: 0.29219012845466735 Epoch: 45 Training loss: 0.14041882872581482 - MAE: 0.2809566443072132 Validation loss : 0.14969545520014232 - MAE: 0.2909081862153574 Epoch: 46 Training loss: 0.140560123026371 - MAE: 0.28191663739835304 Validation loss : 0.14977325167920855 - MAE: 0.291604275381065 Epoch: 47 Training loss: 0.1414617270231247 - MAE: 0.28282842532543745 Validation loss : 0.1506698860062493 - MAE: 0.2930552446484735 Epoch: 48 Training loss: 0.1373920688033104 - MAE: 0.27917131881687096 Validation loss : 0.15224417712953356 - MAE: 0.2934837599479946 Epoch: 49 Training loss: 0.14192317247390748 - MAE: 0.2842585649178634 Validation loss : 0.14885787086354363 - MAE: 0.29192342005218896 Epoch: 50 Training loss: 0.14062062561511993 - MAE: 0.2808979887022064 Validation loss : 0.15054219878382152 - MAE: 0.2924492759311905 Epoch: 51 Training loss: 0.13990722686052323 - MAE: 0.2817114196110173 Validation loss : 0.1506080221798685 - MAE: 0.29247582518131027 Epoch: 52 Training loss: 0.1394276064634323 - MAE: 0.2798831110400087 Validation loss : 0.1490704102648629 - MAE: 0.291844280782309 Epoch: 53 Training loss: 0.14104595750570298 - MAE: 0.28180114573645876 Validation loss : 0.14941397143734825 - MAE: 0.2933880270519655 Epoch: 54 Training loss: 0.139021515250206 - MAE: 0.2793780218867486 Validation loss : 0.1490028988983896 - MAE: 0.29192548589513484 Epoch: 55 Training loss: 0.1409096112847328 - MAE: 0.2816018781967446 Validation loss : 0.1494157827562756 - MAE: 0.2919783378546487 Epoch: 56 Training loss: 0.14150582283735275 - MAE: 0.28397171057514964 Validation loss : 0.1507257032725546 - MAE: 0.29278463527362475 Epoch: 57 Training loss: 0.1412457060813904 - MAE: 0.28194929681996234 Validation loss : 0.1500752924217118 - MAE: 0.291890501089955 Epoch: 58 Training loss: 0.14060731261968612 - MAE: 0.2826944446609872 Validation loss : 0.15134404020176995 - MAE: 0.2921265934844939 Epoch: 59 Training loss: 0.13857035160064698 - MAE: 0.27917573503863885 Validation loss : 0.1502113996280564 - MAE: 0.2916726146711328 Epoch: 60 Training loss: 0.14019066840410233 - MAE: 0.2816340620891049 Validation loss : 0.14821443955103555 - MAE: 0.29117830198396194 Epoch: 61 Training loss: 0.1397022610902786 - MAE: 0.28089572600032964 Validation loss : 0.1513935931854778 - MAE: 0.2917060470497501 Epoch: 62 Training loss: 0.14070956617593766 - MAE: 0.28241979367979564 Validation loss : 0.1501304821835624 - MAE: 0.2912750997306758 Epoch: 63 Training loss: 0.1420893755555153 - MAE: 0.28372013668513324 Validation loss : 0.15038747092088064 - MAE: 0.2920179987433622 Epoch: 64 Training loss: 0.1394691723585129 - MAE: 0.2797968922342149 Validation loss : 0.1498954254719946 - MAE: 0.2930247953251004 Epoch: 65 Training loss: 0.14013640880584716 - MAE: 0.28057332071413116 Validation loss : 0.14924369752407074 - MAE: 0.29117805094765137 Epoch: 66 Training loss: 0.13910346567630769 - MAE: 0.27976148521760813 Validation loss : 0.15365933709674412 - MAE: 0.2938181280395696 Epoch: 67 Training loss: 0.14027394950389863 - MAE: 0.2824978199508478 Validation loss : 0.15143409785297182 - MAE: 0.2915787634959989 Epoch: 68 Training loss: 0.13879832834005357 - MAE: 0.2803663281034843 Validation loss : 0.14806738330258262 - MAE: 0.2899341093266117 Epoch: 69 Training loss: 0.14169045120477677 - MAE: 0.2840059654564348 Validation loss : 0.14932368530167472 - MAE: 0.2908179996636204 Epoch: 70 Training loss: 0.14286602079868316 - MAE: 0.28393754014878414 Validation loss : 0.15445275604724884 - MAE: 0.29465461309717 Epoch: 71 Training loss: 0.1397111463546753 - MAE: 0.2815673786321848 Validation loss : 0.14909260223309198 - MAE: 0.29124006221654514 Epoch: 72 Training loss: 0.13914978861808777 - MAE: 0.27919056115617275 Validation loss : 0.1496173938115438 - MAE: 0.29145609362558494 Epoch: 73 Training loss: 0.13840563476085663 - MAE: 0.2791765280391529 Validation loss : 0.15138710704114702 - MAE: 0.2922445451728484 Epoch: 74 Training loss: 0.13956822723150253 - MAE: 0.2806497885355469 Validation loss : 0.14996923257907233 - MAE: 0.2913337573766691 Epoch: 75 Training loss: 0.13771951258182524 - MAE: 0.2787076577823987 Validation loss : 0.1503291262520684 - MAE: 0.292124609001339 Epoch: 76 Training loss: 0.13808781534433365 - MAE: 0.27866726046665985 Validation loss : 0.14987454232242373 - MAE: 0.29144555209570666 Epoch: 77 Training loss: 0.1389396995306015 - MAE: 0.2791280405088736 Validation loss : 0.1490834727883339 - MAE: 0.29122889859275275 Epoch: 78 Training loss: 0.13928475528955458 - MAE: 0.28034273788540925 Validation loss : 0.14989753978119957 - MAE: 0.2910398033999464 Epoch: 79 Training loss: 0.13676762133836745 - MAE: 0.27751891334295836 Validation loss : 0.15044062005148995 - MAE: 0.29325611713326233 Epoch: 80 Training loss: 0.13998845636844634 - MAE: 0.28041881321085516 Validation loss : 0.14946704192294014 - MAE: 0.293321332550951 Epoch: 81 Training loss: 0.13976493537425994 - MAE: 0.2809897342394576 Validation loss : 0.14930716984801823 - MAE: 0.2915501156778208 Epoch: 82 Training loss: 0.13674694985151292 - MAE: 0.27840960478937093 Validation loss : 0.14946243580844668 - MAE: 0.2906570599625246 Epoch: 83 Training loss: 0.1379085114598274 - MAE: 0.27905460035509927 Validation loss : 0.14950388587183422 - MAE: 0.29112317700446616 Epoch: 84 Training loss: 0.14153614401817322 - MAE: 0.2835935983083774 Validation loss : 0.14871504406134287 - MAE: 0.2905583900515311 Epoch: 85 Training loss: 0.1404113781452179 - MAE: 0.2815505957403188 Validation loss : 0.15186349633667204 - MAE: 0.2935015096363652 Epoch: 86 Training loss: 0.1391930565237999 - MAE: 0.27938914945504073 Validation loss : 0.15015774303012425 - MAE: 0.29175260904666095 Epoch: 87 Training loss: 0.13965154111385344 - MAE: 0.2808129172644999 Validation loss : 0.14953040414386326 - MAE: 0.29217369917685887 Epoch: 88 Training loss: 0.14032937526702882 - MAE: 0.2820906583856728 Validation loss : 0.14954294512669244 - MAE: 0.2906296771253765 Epoch: 89 Training loss: 0.13956192165613174 - MAE: 0.279473451500942 Validation loss : 0.15065168423785102 - MAE: 0.29133002328368285 Epoch: 90 Training loss: 0.14047872692346572 - MAE: 0.2801611359909748 Validation loss : 0.14869611213604608 - MAE: 0.2914427528019158 Epoch: 91 Training loss: 0.14015074878931044 - MAE: 0.28182570327659834 Validation loss : 0.15071149749888313 - MAE: 0.2923330178798639 Epoch: 92 Training loss: 0.14030003398656846 - MAE: 0.2805739854352897 Validation loss : 0.1476295126809014 - MAE: 0.2897238746936595 Epoch: 93 Training loss: 0.1381376227736473 - MAE: 0.27943396254710007 Validation loss : 0.15042445477512148 - MAE: 0.2910839063768687 Epoch: 94 Training loss: 0.13906821191310884 - MAE: 0.27927564341730177 Validation loss : 0.1470845432745086 - MAE: 0.28945548179262404 Epoch: 95 Training loss: 0.13878683894872665 - MAE: 0.27950185756391205 Validation loss : 0.14802363597684437 - MAE: 0.2904005203523502 Epoch: 96 Training loss: 0.1400198295712471 - MAE: 0.28227933824874535 Validation loss : 0.14741118169493145 - MAE: 0.2886919057055488 Epoch: 97 Training loss: 0.14008406758308412 - MAE: 0.2820675306746204 Validation loss : 0.1483982594476806 - MAE: 0.2895298978938563 Epoch: 98 Training loss: 0.13766043066978453 - MAE: 0.27798967107381206 Validation loss : 0.1500189420249727 - MAE: 0.29136640929113017 Epoch: 99 Training loss: 0.14085446119308473 - MAE: 0.28138468637744457 Validation loss : 0.147224517332183 - MAE: 0.28914512721321917 Epoch: 100 Training loss: 0.140465350151062 - MAE: 0.28220289946670496 Validation loss : 0.14812367161115012 - MAE: 0.289046826605416 Epoch: 101 Training loss: 0.13990859061479569 - MAE: 0.2795642386535082 Validation loss : 0.1493135475450092 - MAE: 0.29040059011073405 Epoch: 102 Training loss: 0.14165273666381836 - MAE: 0.28245179865396797 Validation loss : 0.14880059079991448 - MAE: 0.29071234075534924 Epoch: 103 Training loss: 0.1388886308670044 - MAE: 0.279369814576424 Validation loss : 0.1476850998070505 - MAE: 0.2903485271171268 Epoch: 104 Training loss: 0.13909724473953247 - MAE: 0.28083273552929844 Validation loss : 0.14797810382313198 - MAE: 0.28941232666977995 Epoch: 105 Training loss: 0.1393595016002655 - MAE: 0.2783428221019483 Validation loss : 0.14831788175635868 - MAE: 0.2905713458472301 Epoch: 106 Training loss: 0.1390196332335472 - MAE: 0.27999429469540527 Validation loss : 0.1493500214484003 - MAE: 0.29255232062385816 Epoch: 107 Training loss: 0.1404520383477211 - MAE: 0.28146161503782474 Validation loss : 0.14878038648102018 - MAE: 0.2906487705342995 Epoch: 108 Training loss: 0.14004070401191712 - MAE: 0.28177979907550266 Validation loss : 0.14799664749039543 - MAE: 0.28897706210829416 Epoch: 109 Training loss: 0.1383381900191307 - MAE: 0.2784476944130519 Validation loss : 0.1490860970483886 - MAE: 0.2903748192836748 Epoch: 110 Training loss: 0.14010255575180053 - MAE: 0.2814791462622622 Validation loss : 0.14862897247076035 - MAE: 0.2912308111745315 Epoch: 111 Training loss: 0.1402219730615616 - MAE: 0.2815668657801355 Validation loss : 0.14855213463306427 - MAE: 0.2909928236013806 Epoch: 112 Training loss: 0.13920035630464553 - MAE: 0.281543453418757 Validation loss : 0.14755674368805355 - MAE: 0.2900560617682023 Epoch: 113 Training loss: 0.1400330564379692 - MAE: 0.2802442792672325 Validation loss : 0.14960995233721203 - MAE: 0.29297903682515697 Epoch: 114 Training loss: 0.13953691601753235 - MAE: 0.2798329911916557 Validation loss : 0.14831144445472294 - MAE: 0.2901465696869431 Epoch: 115 Training loss: 0.13693355649709701 - MAE: 0.2773904079655369 Validation loss : 0.14724988407558864 - MAE: 0.2901089629035545 Epoch: 116 Training loss: 0.13829326421022414 - MAE: 0.28039887425826104 Validation loss : 0.14879283474551308 - MAE: 0.29064504580363937 Epoch: 117 Training loss: 0.14090506106615067 - MAE: 0.2826000565784893 Validation loss : 0.1477228750785192 - MAE: 0.2903357764542001 Epoch: 118 Training loss: 0.1429896029829979 - MAE: 0.2845150458178373 Validation loss : 0.1520556633671125 - MAE: 0.2927788288379692 Epoch: 119 Training loss: 0.14179373949766158 - MAE: 0.28321784029712616 Validation loss : 0.1478977551062902 - MAE: 0.28899535825863276 Epoch: 120 Training loss: 0.13644426107406615 - MAE: 0.277860572559567 Validation loss : 0.1487949498825603 - MAE: 0.289971918515328 Epoch: 121 Training loss: 0.13941168040037155 - MAE: 0.28066886387000756 Validation loss : 0.14904049038887024 - MAE: 0.2892359740394001 Epoch: 122 Training loss: 0.13944829851388932 - MAE: 0.28015464875184926 Validation loss : 0.14853730301062265 - MAE: 0.28979281078938246 Epoch: 123 Training loss: 0.13739350140094758 - MAE: 0.27918069067817697 Validation loss : 0.14981382257408565 - MAE: 0.29127630466726984 Epoch: 124 Training loss: 0.13999282330274582 - MAE: 0.2798265373231701 Validation loss : 0.14869198782576454 - MAE: 0.29036692179389667 Epoch: 125 Training loss: 0.14108887642621995 - MAE: 0.28223551065679303 Validation loss : 0.1493358156747288 - MAE: 0.29129564350186565 Epoch: 126 Training loss: 0.13976523160934448 - MAE: 0.2817742714626282 Validation loss : 0.14911426686578327 - MAE: 0.29106398685631213 Epoch: 127 Training loss: 0.14017222493886947 - MAE: 0.28179540480169774 Validation loss : 0.14842878530422846 - MAE: 0.2901467176758952 Epoch: 128 Training loss: 0.1406565561890602 - MAE: 0.28163913944949737 Validation loss : 0.1479131720132298 - MAE: 0.28919518276952677 Epoch: 129 Training loss: 0.13942452102899552 - MAE: 0.2793156197663516 Validation loss : 0.1511403951379988 - MAE: 0.29067412905683 Epoch: 130 Training loss: 0.14031505227088928 - MAE: 0.2804210714866914 Validation loss : 0.14765288515223396 - MAE: 0.288500744825017 Epoch: 131 Training loss: 0.13649964869022368 - MAE: 0.2774726677017398 Validation loss : 0.14891337934467527 - MAE: 0.2914018373625627 Epoch: 132 Training loss: 0.1405644327402115 - MAE: 0.2804009316942137 Validation loss : 0.1483287529812919 - MAE: 0.2907619964928693 Epoch: 133 Training loss: 0.14118699222803116 - MAE: 0.2819625654924857 Validation loss : 0.15040278683106104 - MAE: 0.2910529233222426 Epoch: 134 Training loss: 0.14081039100885392 - MAE: 0.28238484691454196 Validation loss : 0.14994350406858656 - MAE: 0.2906808160616756 Epoch: 135 Training loss: 0.13862200796604157 - MAE: 0.27899754217163214 Validation loss : 0.14944083078040016 - MAE: 0.2906151099314951 Epoch: 136 Training loss: 0.13931725412607193 - MAE: 0.28030006822907155 Validation loss : 0.1495074381430944 - MAE: 0.29074428772408845 Epoch: 137 Training loss: 0.1403974002599716 - MAE: 0.2805740067883291 Validation loss : 0.148383229970932 - MAE: 0.29007782636865753 Epoch: 138 Training loss: 0.1385694068670273 - MAE: 0.27983454104510636 Validation loss : 0.1512098577287462 - MAE: 0.2919661208163271 Epoch: 139 Training loss: 0.1404304113984108 - MAE: 0.28148055927808185 Validation loss : 0.1489351209666994 - MAE: 0.29030302398224317 Epoch: 140 Training loss: 0.1390806904435158 - MAE: 0.27891810870400857 Validation loss : 0.1524661953250567 - MAE: 0.29273273639235975 Epoch: 141 Training loss: 0.13796700328588485 - MAE: 0.27966280744881894 Validation loss : 0.15020373629199135 - MAE: 0.2912314458069829 Epoch: 142 Training loss: 0.14071892708539963 - MAE: 0.2816539302440052 Validation loss : 0.15058360497156778 - MAE: 0.2914712531041945 Epoch: 143 Training loss: 0.14057636469602586 - MAE: 0.28242232736529516 Validation loss : 0.1506333847840627 - MAE: 0.29140815832807077 Epoch: 144 Training loss: 0.13925146371126174 - MAE: 0.2799103959499411 Validation loss : 0.14928915599981943 - MAE: 0.2897573463488531 Epoch: 145 Training loss: 0.1398252120614052 - MAE: 0.28164205446635043 Validation loss : 0.14851086835066477 - MAE: 0.29021297413113184 Epoch: 146 Training loss: 0.13817338794469833 - MAE: 0.27937391482438806 Validation loss : 0.1498860807882415 - MAE: 0.29185029747881874 Epoch: 147 Training loss: 0.14039424389600755 - MAE: 0.28106582531667956 Validation loss : 0.14853160911136204 - MAE: 0.29055607567586406 Epoch: 148 Training loss: 0.14145232886075973 - MAE: 0.2822118953799359 Validation loss : 0.15043112718396717 - MAE: 0.2923702620366525 Epoch: 149 Training loss: 0.14144912719726563 - MAE: 0.281877632162409 Validation loss : 0.1527728090683619 - MAE: 0.2934056300021226 Epoch: 150 Training loss: 0.14076999455690384 - MAE: 0.282006950149505 Validation loss : 0.15191659910811317 - MAE: 0.29212817923698803 Epoch: 151 Training loss: 0.14104728668928146 - MAE: 0.282059518005827 Validation loss : 0.15123689423004785 - MAE: 0.29217216756303294 Epoch: 152 Training loss: 0.14049483269453047 - MAE: 0.2816058638889221 Validation loss : 0.14788557837406793 - MAE: 0.29005179357861427 Epoch: 153 Training loss: 0.13666647017002106 - MAE: 0.276851346022365 Validation loss : 0.1481863541735543 - MAE: 0.2906076278179208 Epoch: 154 Training loss: 0.1380725646018982 - MAE: 0.27842498943152266 Validation loss : 0.14886879755391014 - MAE: 0.29055297838102373 Epoch: 155 Training loss: 0.13862104922533036 - MAE: 0.2793241343310853 Validation loss : 0.14891281558407676 - MAE: 0.29119474966859027 Epoch: 156 Training loss: 0.13700734436511994 - MAE: 0.278056366301915 Validation loss : 0.1506768903798527 - MAE: 0.2922085560624545 Epoch: 157 Training loss: 0.13953308165073394 - MAE: 0.281015862464819 Validation loss : 0.1483795220653216 - MAE: 0.29058860133117714 Epoch: 158 Training loss: 0.14014002352952956 - MAE: 0.28094309193212397 Validation loss : 0.1488533839583397 - MAE: 0.2904079557435194 Epoch: 159 Training loss: 0.1392805379629135 - MAE: 0.280401866801906 Validation loss : 0.152536627319124 - MAE: 0.29208377007193054 Epoch: 160 Training loss: 0.13968930572271346 - MAE: 0.2801028720194759 Validation loss : 0.15070715049902597 - MAE: 0.2915361243575389 Epoch: 161 Training loss: 0.1379876920580864 - MAE: 0.2779706198029852 Validation loss : 0.14963952700297037 - MAE: 0.29140363323705737 Epoch: 162 Training loss: 0.13882436752319335 - MAE: 0.2803142060072604 Validation loss : 0.14831043117576176 - MAE: 0.29090746702578535 Epoch: 163 Training loss: 0.14012926697731018 - MAE: 0.2826213295889931 Validation loss : 0.15037808484501308 - MAE: 0.2921038937559237 Epoch: 164 Training loss: 0.13816417813301085 - MAE: 0.27881172247852454 Validation loss : 0.14971058236228096 - MAE: 0.29022949264590414 Epoch: 165 Training loss: 0.14038401901721953 - MAE: 0.28114290350520105 Validation loss : 0.14926749053928587 - MAE: 0.2900267579172161 Epoch: 166 Training loss: 0.14115286707878114 - MAE: 0.2812040498053005 Validation loss : 0.15122641126314798 - MAE: 0.29175891756760863 Epoch: 167 Training loss: 0.13707010865211486 - MAE: 0.2777921438305016 Validation loss : 0.150983441207144 - MAE: 0.29142422296616527 Epoch: 168 Training loss: 0.13876796901226043 - MAE: 0.2802754449333229 Validation loss : 0.14823240455653933 - MAE: 0.2911944734775949 Epoch: 169 Training loss: 0.1400000700354576 - MAE: 0.27962448113148264 Validation loss : 0.15024596121576098 - MAE: 0.2927376945330595 Epoch: 170 Training loss: 0.13957252651453017 - MAE: 0.27999991303737254 Validation loss : 0.15100336488750246 - MAE: 0.29298232371488797 Epoch: 171 Training loss: 0.1379392111301422 - MAE: 0.28030924084273134 Validation loss : 0.1500499521692594 - MAE: 0.29190070752057323 Epoch: 172 Training loss: 0.13905995577573776 - MAE: 0.27959100752691673 Validation loss : 0.1501365229487419 - MAE: 0.291390080732848 Epoch: 173 Training loss: 0.13984983891248703 - MAE: 0.28052682384061434 Validation loss : 0.15079624785317314 - MAE: 0.2914017756619142 Epoch: 174 Training loss: 0.13995950669050217 - MAE: 0.28073402267442327 Validation loss : 0.15117037296295166 - MAE: 0.2913482358293418 Epoch: 175 Training loss: 0.13972256630659102 - MAE: 0.2810884447912753 Validation loss : 0.1483582788043552 - MAE: 0.2891574760822575 Epoch: 176 Training loss: 0.1396074891090393 - MAE: 0.2802075005571549 Validation loss : 0.15012793242931366 - MAE: 0.29115621681530746 Epoch: 177 Training loss: 0.14028172552585602 - MAE: 0.2809140783358514 Validation loss : 0.15086209111743504 - MAE: 0.2916531135350998 Epoch: 178 Training loss: 0.13791372925043105 - MAE: 0.27898197003406183 Validation loss : 0.15098679976330864 - MAE: 0.2925168924383861 Epoch: 179 Training loss: 0.1382387799024582 - MAE: 0.27801012015838344 Validation loss : 0.1535450733370251 - MAE: 0.29484870415135217 Epoch: 180 Training loss: 0.1407102045416832 - MAE: 0.28173970451231156 Validation loss : 0.1496893283393648 - MAE: 0.2922256774422862 Epoch: 181 Training loss: 0.13767244040966034 - MAE: 0.27825487061974347 Validation loss : 0.14896611786550945 - MAE: 0.2912224345710289 Epoch: 182 Training loss: 0.1394115325808525 - MAE: 0.28027065459265726 Validation loss : 0.14866195287969378 - MAE: 0.29134036191775803 Epoch: 183 Training loss: 0.13798033624887465 - MAE: 0.27942216029668876 Validation loss : 0.1484985484017266 - MAE: 0.2912727374193501 Epoch: 184 Training loss: 0.1377663853764534 - MAE: 0.2785826649941721 Validation loss : 0.1484346944424841 - MAE: 0.2905791137891933 Epoch: 185 Training loss: 0.13849070370197297 - MAE: 0.27878410506584883 Validation loss : 0.15334946744971806 - MAE: 0.2941069864986494 Epoch: 186 Training loss: 0.1423814630508423 - MAE: 0.28398425886455253 Validation loss : 0.14981804374191496 - MAE: 0.2910886664842828 Epoch: 187 Training loss: 0.13906280905008317 - MAE: 0.2796231689952685 Validation loss : 0.14884889953666264 - MAE: 0.29072980795626585 Epoch: 188 Training loss: 0.13866673171520233 - MAE: 0.27987766929421665 Validation loss : 0.15112879872322083 - MAE: 0.29146268896748756 Epoch: 189 Training loss: 0.13786116987466812 - MAE: 0.2796589432405557 Validation loss : 0.1473951761921247 - MAE: 0.2899062293933996 Epoch: 190 Training loss: 0.1392700281739235 - MAE: 0.28174620813821255 Validation loss : 0.14669430669811037 - MAE: 0.2889526807976215 Epoch: 191 Training loss: 0.1394435104727745 - MAE: 0.2802861075304156 Validation loss : 0.1477406405740314 - MAE: 0.2888190542998886 Epoch: 192 Training loss: 0.14042726933956146 - MAE: 0.28141030296388986 Validation loss : 0.15162903236018288 - MAE: 0.29180724059413954 Epoch: 193 Training loss: 0.13837426751852036 - MAE: 0.27898095057570355 Validation loss : 0.14722405870755514 - MAE: 0.2892951285523225 Epoch: 194 Training loss: 0.13956972509622573 - MAE: 0.2816792570432636 Validation loss : 0.15053565055131912 - MAE: 0.291487144598652 Epoch: 195 Training loss: 0.13988079249858856 - MAE: 0.28193311766088014 Validation loss : 0.1497278105881479 - MAE: 0.2906707147216234 Epoch: 196 Training loss: 0.13724557787179947 - MAE: 0.2774827554632601 Validation loss : 0.14994101391898262 - MAE: 0.2910469203395106 Epoch: 197 Training loss: 0.13723361849784851 - MAE: 0.27802080726455736 Validation loss : 0.14813513474331963 - MAE: 0.2891774682253844 Epoch: 198 Training loss: 0.13782596826553345 - MAE: 0.27820030438241256