File size: 25,360 Bytes
a9614e2 1c266c2 573d184 138c452 085fc0e a8e4a32 60030c6 43aca8f bced6f4 3c0ba3e 05cb146 2755fdc 6d43361 9ba6b1c bc4b6fa fd66402 8ac5557 0a5b886 7582378 7943a4b 1d207e4 fc0b9fb 778b2ee d8f234d b8f30b4 07d6b01 b1e9821 5d63be6 c240e23 7fd6a30 753d906 7988c17 f9cc4b4 2563dd5 3b32d9d 78c220f 13dd3c9 3645ffc a67ad3c 1791d1d 417295b 9d01042 4692ce2 ee702fe 9ad9b31 8a55390 af3dc92 255d975 b5c59b2 6291ba6 910096e f1cb193 5ecc42f f52fd05 4aa4325 5b2f689 0ca188f 90fba87 3b6adf1 d7b4484 b3d9166 822020b 5f1eb05 5a03441 36bc79e ff030dd 927b3c1 64f46f0 749183e 2c6fb37 bff4a2d 46697fc 0413f7a 5ee6dd5 419fd28 eb700da 64c500b d0056fd 66982b8 1d41cd7 1c356ba d930327 fcc9b1f 1399e97 04ba03c ed8b3f4 025040b 13b93d7 d26ef5a a37fc5f cf6faf1 86a6afb 40f3e9f 7cbbdc1 f5369b2 63b151c eaadc50 6354a95 0d8ed18 8182faf f9e751c cb1f440 8a3fa68 21062cf 7ce2a2c 1f989a1 3d97e81 23665ab 57ada11 4a041ea 91a97e6 f4e0960 42a05eb 214e4e3 a4c8ea3 1d48055 55af590 864f0d5 d12921d b7fd40a e4b2e16 35063ea c287647 9cc161b 16df9c1 8c239b2 e1ffb75 f067b6b 5707b55 06dbd2d ba89237 748bfcb ce6bb9d a0c316e 0e41bee a7c336c 284862e e6bfda7 ff1b253 7f85e3c 39360cd bf774fd 751781f 13f6f78 a14234f 2e1113b 26dbf57 48ecfcf c37f189 5e1b947 76ec1ac bc66fa8 71741a1 6fe9bdc cba9e16 e6484a9 e9f9dc2 e2560d5 94d08de ab2ab8b 306c777 b0daf82 7161dc2 ec8b412 6e7f355 1dbcf89 74c5bce 8f28059 cd5e1ba 50d54b4 e3854b6 47f5e22 0e66be8 0596de2 67a7610 eacd439 610aa7d 76212f8 5a43060 94586e7 6fae1c1 8ce35ef de84ad0 71f0a6e 7988ea4 2ce4e82 0b0e4ff e51065c c85d65b be9081c d27e527 4fe4ad7 edff910 8f89e59 f773786 0c18394 e5713c6 4257128 587cfee 2855675 c951d8f 26d82b2 668d135 cf661ef 5d6ba54 e07588e 0bf1606 7ba5a4f 5ccf23e 04c8203 3ce5140 0142e2d aa8c5f1 14e679e 5e114da 3fa5453 2d659e3 d996cf0 56ae8c1 0ff9e6e cef854e db07dce b013ee1 af9d814 c9d4c79 270f1db 0a8d8f5 757246c b4908df fb70983 1254ee4 9bd0527 adef98d 7fdaae7 f5aec4e 63e1b2c d1c2ef2 cdfafca d98eaa9 d8b886a 411270c 8f14fa0 0b4e95a 5fd9305 1c6a829 5359fa5 7e68e76 a3d1fa4 1d756fc 9789da4 b510afb 31e8acb 6ffc0ea f558ed5 7989194 28131e9 55d92e6 5570ada c30434c b7bf10c a50cfb9 328ddd3 ca5360b df0a334 b9f3062 c20973c 3b2d473 6279f3b 2192d9b 6c052e1 35a62f3 9cd94e6 cea3250 1b7ab72 8c30aff ba362ed b20f700 dd5c50f 4eb1ab2 9935079 5b1f10c fd34687 a095ee0 7117070 f93d704 222ad96 82126f9 bb5e24a e96b271 7fcc0fd 36b0b74 ea78722 9584d30 72d3b07 de2672e 7faba00 e0bb648 10e4a00 0d74544 448c740 19de6ba 3f0b83f 14fcabb b5a78e6 fcb1ba2 bc901e2 0c1ecf9 21aa32a 6650600 1881e39 186a8b7 146d0b6 47ff58e 3fab5ae 3c7dcd2 c59732a c83ce85 db31097 a70bbcd a365d1d a53a7e5 e66e540 915d84c 3175739 0703d6e 96dd1ef e9e6722 3056e8f e51210c ddfed9b fc48cdd 84ddc15 99b7715 f63c516 a9a3095 5676aaa 134be0e c73423c 2d505db f94c81c 84df553 75845e8 82ba38b 171c0bb 8aee9c4 ec41b2e 5ee5721 5330555 a2e9cde 37a56dd e3f480a f1adf41 fd7181d 02a622c cb5dfef 6607028 eb07595 9109f82 fc45727 fcdf3fb 0bba515 c1a350b 00bec49 8415958 c9e290a 0185638 e8b06c1 90b2b0d 486ccb8 6cffbbe 0e1162e 78a40bb 8f538a2 040d1b9 bd34f15 4377c2a 856a1c9 5320034 93e268a 9766d65 db83950 497cace d0badb0 a536fb2 396420c c61131b abb5dfd 60c0337 687f13d 9b105ab 5f9a66b 1c0ed1e e0896ef f6401d3 809478d 5351220 198c64a 5774464 f87e23c 1f0928c d8fdcc1 f0a5160 2095131 709aa53 45ff62e 7a35414 a770ba7 e82fd4e 8ca9206 01a2885 8177a48 0381edc 80bca49 357e2f3 06317b7 1712fb6 0186558 3707add 8263364 08bcb83 b7c6653 09ec5d4 94a4783 b07944c a863d27 e57e26d 5878ae0 b3d82a9 17a8b64 966a3f4 fdf7af1 c5a7da2 5d3cf93 96c941a 5113641 e7e83b7 de68187 ec2f96c dce162e 42dceef 880d82c ba963f3 45e8f42 76728d2 659aae1 10c3671 6ffae2f 1ac1bdc 979edaf e359218 eb0db72 e87bcc0 5b695c4 e9a7682 4a9719e 2ac2a49 5397188 726de91 b9a2c9e 9370705 9b93aa0 a3466a0 ecf9f75 2f82651 33a8b37 da5fa69 9c43a33 49d927d b97f761 9ee5878 6da5129 dddddad da35d46 8c0e081 07d6c80 4888b06 5992514 7f5a03e d8ccce8 a16965d 9efc5ff 995c064 a9e4d43 3aa3a40 2fee259 56ad833 fe9559e 6d6711e f54804f f46c363 d5ef9b9 7f9edc3 ac0a5bd 3a52a27 5b3592a 12777d5 01daf53 0e81065 0539c7f faf1d5b f8afb52 839ff59 699f4e8 c054eb4 6a81f5b 4560197 22daea2 c4f289b 63809dd c9f0acc 95ed9ee 7a34fd8 b3d9418 d0d4876 e1c09e6 6d76ed5 afef343 0406e2f 85384d2 b484e26 16898d3 09de62b 97c5805 a3d18b0 0ca7896 f130240 126db89 f1c7481 0585093 a0d7d1f 88c3644 a6f1dca 7dd3553 8f20afd 84ce014 c46cc4f 57f3fc6 05cd87c 136fad9 23259b1 c9ff948 eb5f148 cbb02c8 9d6d79a 10cc508 2cef535 20d7799 3b45ca6 120811f 03ef9b2 95d2eab f3bd43d 0560de0 2b05bc0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 |
Started at: 21:34:05 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.32175946950912476 - MAE: 0.4437903503551826 Validation loss : 0.19621961481041378 - MAE: 0.3388283822410594 Epoch: 1 Training loss: 0.21072698533535003 - MAE: 0.35116530289867337 Validation loss : 0.1876767095592287 - MAE: 0.33628140968919695 Epoch: 2 Training loss: 0.1893942326307297 - MAE: 0.3312522516218666 Validation loss : 0.1748287214173211 - MAE: 0.3213797447248068 Epoch: 3 Training loss: 0.17969159841537474 - MAE: 0.3215382240664255 Validation loss : 0.1697687026527193 - MAE: 0.31446870224950263 Epoch: 4 Training loss: 0.17220519423484804 - MAE: 0.3129526796594085 Validation loss : 0.16811848680178323 - MAE: 0.3117262145282266 Epoch: 5 Training loss: 0.17296144247055054 - MAE: 0.31494873624825814 Validation loss : 0.16849000917540657 - MAE: 0.31104514179599985 Epoch: 6 Training loss: 0.16641035348176955 - MAE: 0.30768475494362546 Validation loss : 0.16421516074074638 - MAE: 0.30715220958263423 Epoch: 7 Training loss: 0.16057054400444032 - MAE: 0.30329494898002457 Validation loss : 0.161027698053254 - MAE: 0.304297376785587 Epoch: 8 Training loss: 0.1566170272231102 - MAE: 0.29957014870206655 Validation loss : 0.15982638630602095 - MAE: 0.30397747682822174 Epoch: 9 Training loss: 0.15673983812332154 - MAE: 0.3000540458404174 Validation loss : 0.159184659520785 - MAE: 0.3035235378542429 Epoch: 10 Training loss: 0.15630604147911073 - MAE: 0.29690365842430627 Validation loss : 0.15921704471111298 - MAE: 0.30240467396157955 Epoch: 11 Training loss: 0.15595020622015 - MAE: 0.29754135005638765 Validation loss : 0.15686986181471083 - MAE: 0.30119996351152656 Epoch: 12 Training loss: 0.15274528950452804 - MAE: 0.2944161972508913 Validation loss : 0.15652166141404045 - MAE: 0.3012713923501961 Epoch: 13 Training loss: 0.15205995708703995 - MAE: 0.29424324063629004 Validation loss : 0.1575678288936615 - MAE: 0.3004794443503115 Epoch: 14 Training loss: 0.15087180227041244 - MAE: 0.292504579327589 Validation loss : 0.15574459234873453 - MAE: 0.29932356111019714 Epoch: 15 Training loss: 0.1518820345401764 - MAE: 0.2940904971897683 Validation loss : 0.15581322544150883 - MAE: 0.30048684662514935 Epoch: 16 Training loss: 0.14916340589523316 - MAE: 0.2900369708605816 Validation loss : 0.15502946575482687 - MAE: 0.2986623058669649 Epoch: 17 Training loss: 0.14969733864068985 - MAE: 0.2916869417468108 Validation loss : 0.15631223718325296 - MAE: 0.2994015598567933 Epoch: 18 Training loss: 0.14621972769498826 - MAE: 0.2875086269286061 Validation loss : 0.1557358337773217 - MAE: 0.29931970436403404 Epoch: 19 Training loss: 0.14798570185899734 - MAE: 0.29024787778757843 Validation loss : 0.15424930387073094 - MAE: 0.2984811820958494 Epoch: 20 Training loss: 0.14769238144159316 - MAE: 0.2896109423923894 Validation loss : 0.1536422868569692 - MAE: 0.2987445844262027 Epoch: 21 Training loss: 0.14361368536949157 - MAE: 0.2842206176667335 Validation loss : 0.1533755792511834 - MAE: 0.29642898867488277 Epoch: 22 Training loss: 0.14456430345773696 - MAE: 0.2860957867005398 Validation loss : 0.1542035871081882 - MAE: 0.2971775973100257 Epoch: 23 Training loss: 0.1426533755660057 - MAE: 0.28512021628758083 Validation loss : 0.15245803362793392 - MAE: 0.2961228783995986 Epoch: 24 Training loss: 0.14542177826166153 - MAE: 0.2855342243309425 Validation loss : 0.15294104317824045 - MAE: 0.29643245601047447 Epoch: 25 Training loss: 0.14522234290838243 - MAE: 0.2851345876886818 Validation loss : 0.15342944694889915 - MAE: 0.2958033843008907 Epoch: 26 Training loss: 0.14386597275733948 - MAE: 0.2838243646140143 Validation loss : 0.1515038808186849 - MAE: 0.2936464496284943 Epoch: 27 Training loss: 0.1415349954366684 - MAE: 0.2810586437104536 Validation loss : 0.1507236527072059 - MAE: 0.29409404478735646 Epoch: 28 Training loss: 0.14115683376789093 - MAE: 0.2815221038197439 Validation loss : 0.15061336921321022 - MAE: 0.29280129784932213 Epoch: 29 Training loss: 0.14186541587114335 - MAE: 0.28239653007887255 Validation loss : 0.15151139597098032 - MAE: 0.2953867482779256 Epoch: 30 Training loss: 0.14276256740093232 - MAE: 0.283208699006944 Validation loss : 0.1504600097735723 - MAE: 0.29276670315056585 Epoch: 31 Training loss: 0.14241950929164887 - MAE: 0.28277882078383293 Validation loss : 0.1505332812666893 - MAE: 0.29293145394981956 Epoch: 32 Training loss: 0.14204519629478454 - MAE: 0.28192363607736504 Validation loss : 0.15044088496102226 - MAE: 0.2929076596067834 Epoch: 33 Training loss: 0.13920597046613692 - MAE: 0.2793298976614221 Validation loss : 0.15006180769867367 - MAE: 0.29292432343849906 Epoch: 34 Training loss: 0.1410813584923744 - MAE: 0.2815379053911258 Validation loss : 0.14930015967951882 - MAE: 0.2933547268525188 Epoch: 35 Training loss: 0.14150760889053346 - MAE: 0.2832031330097777 Validation loss : 0.15295125875208113 - MAE: 0.2936935655701593 Epoch: 36 Training loss: 0.1405733221769333 - MAE: 0.2821643593441243 Validation loss : 0.14867112785577774 - MAE: 0.2921335742814868 Epoch: 37 Training loss: 0.14206359952688216 - MAE: 0.28261241361934286 Validation loss : 0.14988169405195448 - MAE: 0.2927604772074144 Epoch: 38 Training loss: 0.14064243495464324 - MAE: 0.28185439579063915 Validation loss : 0.1515509072277281 - MAE: 0.29319069832718325 Epoch: 39 Training loss: 0.14297345608472825 - MAE: 0.282847817609895 Validation loss : 0.15042786465750801 - MAE: 0.29282722784595605 Epoch: 40 Training loss: 0.13990240722894667 - MAE: 0.27964537481988466 Validation loss : 0.14972211172183356 - MAE: 0.29260355981671565 Epoch: 41 Training loss: 0.13921964198350906 - MAE: 0.28078253547199195 Validation loss : 0.14965214083592096 - MAE: 0.2924700199853063 Epoch: 42 Training loss: 0.1393338233232498 - MAE: 0.28024463401390987 Validation loss : 0.1507701873779297 - MAE: 0.29355734742186085 Epoch: 43 Training loss: 0.1382669734954834 - MAE: 0.27992590956331337 Validation loss : 0.14967524343066746 - MAE: 0.29149855699855537 Epoch: 44 Training loss: 0.14029370576143266 - MAE: 0.28265328374425863 Validation loss : 0.1491319098406368 - MAE: 0.29219012845466735 Epoch: 45 Training loss: 0.14041882872581482 - MAE: 0.2809566443072132 Validation loss : 0.14969545520014232 - MAE: 0.2909081862153574 Epoch: 46 Training loss: 0.140560123026371 - MAE: 0.28191663739835304 Validation loss : 0.14977325167920855 - MAE: 0.291604275381065 Epoch: 47 Training loss: 0.1414617270231247 - MAE: 0.28282842532543745 Validation loss : 0.1506698860062493 - MAE: 0.2930552446484735 Epoch: 48 Training loss: 0.1373920688033104 - MAE: 0.27917131881687096 Validation loss : 0.15224417712953356 - MAE: 0.2934837599479946 Epoch: 49 Training loss: 0.14192317247390748 - MAE: 0.2842585649178634 Validation loss : 0.14885787086354363 - MAE: 0.29192342005218896 Epoch: 50 Training loss: 0.14062062561511993 - MAE: 0.2808979887022064 Validation loss : 0.15054219878382152 - MAE: 0.2924492759311905 Epoch: 51 Training loss: 0.13990722686052323 - MAE: 0.2817114196110173 Validation loss : 0.1506080221798685 - MAE: 0.29247582518131027 Epoch: 52 Training loss: 0.1394276064634323 - MAE: 0.2798831110400087 Validation loss : 0.1490704102648629 - MAE: 0.291844280782309 Epoch: 53 Training loss: 0.14104595750570298 - MAE: 0.28180114573645876 Validation loss : 0.14941397143734825 - MAE: 0.2933880270519655 Epoch: 54 Training loss: 0.139021515250206 - MAE: 0.2793780218867486 Validation loss : 0.1490028988983896 - MAE: 0.29192548589513484 Epoch: 55 Training loss: 0.1409096112847328 - MAE: 0.2816018781967446 Validation loss : 0.1494157827562756 - MAE: 0.2919783378546487 Epoch: 56 Training loss: 0.14150582283735275 - MAE: 0.28397171057514964 Validation loss : 0.1507257032725546 - MAE: 0.29278463527362475 Epoch: 57 Training loss: 0.1412457060813904 - MAE: 0.28194929681996234 Validation loss : 0.1500752924217118 - MAE: 0.291890501089955 Epoch: 58 Training loss: 0.14060731261968612 - MAE: 0.2826944446609872 Validation loss : 0.15134404020176995 - MAE: 0.2921265934844939 Epoch: 59 Training loss: 0.13857035160064698 - MAE: 0.27917573503863885 Validation loss : 0.1502113996280564 - MAE: 0.2916726146711328 Epoch: 60 Training loss: 0.14019066840410233 - MAE: 0.2816340620891049 Validation loss : 0.14821443955103555 - MAE: 0.29117830198396194 Epoch: 61 Training loss: 0.1397022610902786 - MAE: 0.28089572600032964 Validation loss : 0.1513935931854778 - MAE: 0.2917060470497501 Epoch: 62 Training loss: 0.14070956617593766 - MAE: 0.28241979367979564 Validation loss : 0.1501304821835624 - MAE: 0.2912750997306758 Epoch: 63 Training loss: 0.1420893755555153 - MAE: 0.28372013668513324 Validation loss : 0.15038747092088064 - MAE: 0.2920179987433622 Epoch: 64 Training loss: 0.1394691723585129 - MAE: 0.2797968922342149 Validation loss : 0.1498954254719946 - MAE: 0.2930247953251004 Epoch: 65 Training loss: 0.14013640880584716 - MAE: 0.28057332071413116 Validation loss : 0.14924369752407074 - MAE: 0.29117805094765137 Epoch: 66 Training loss: 0.13910346567630769 - MAE: 0.27976148521760813 Validation loss : 0.15365933709674412 - MAE: 0.2938181280395696 Epoch: 67 Training loss: 0.14027394950389863 - MAE: 0.2824978199508478 Validation loss : 0.15143409785297182 - MAE: 0.2915787634959989 Epoch: 68 Training loss: 0.13879832834005357 - MAE: 0.2803663281034843 Validation loss : 0.14806738330258262 - MAE: 0.2899341093266117 Epoch: 69 Training loss: 0.14169045120477677 - MAE: 0.2840059654564348 Validation loss : 0.14932368530167472 - MAE: 0.2908179996636204 Epoch: 70 Training loss: 0.14286602079868316 - MAE: 0.28393754014878414 Validation loss : 0.15445275604724884 - MAE: 0.29465461309717 Epoch: 71 Training loss: 0.1397111463546753 - MAE: 0.2815673786321848 Validation loss : 0.14909260223309198 - MAE: 0.29124006221654514 Epoch: 72 Training loss: 0.13914978861808777 - MAE: 0.27919056115617275 Validation loss : 0.1496173938115438 - MAE: 0.29145609362558494 Epoch: 73 Training loss: 0.13840563476085663 - MAE: 0.2791765280391529 Validation loss : 0.15138710704114702 - MAE: 0.2922445451728484 Epoch: 74 Training loss: 0.13956822723150253 - MAE: 0.2806497885355469 Validation loss : 0.14996923257907233 - MAE: 0.2913337573766691 Epoch: 75 Training loss: 0.13771951258182524 - MAE: 0.2787076577823987 Validation loss : 0.1503291262520684 - MAE: 0.292124609001339 Epoch: 76 Training loss: 0.13808781534433365 - MAE: 0.27866726046665985 Validation loss : 0.14987454232242373 - MAE: 0.29144555209570666 Epoch: 77 Training loss: 0.1389396995306015 - MAE: 0.2791280405088736 Validation loss : 0.1490834727883339 - MAE: 0.29122889859275275 Epoch: 78 Training loss: 0.13928475528955458 - MAE: 0.28034273788540925 Validation loss : 0.14989753978119957 - MAE: 0.2910398033999464 Epoch: 79 Training loss: 0.13676762133836745 - MAE: 0.27751891334295836 Validation loss : 0.15044062005148995 - MAE: 0.29325611713326233 Epoch: 80 Training loss: 0.13998845636844634 - MAE: 0.28041881321085516 Validation loss : 0.14946704192294014 - MAE: 0.293321332550951 Epoch: 81 Training loss: 0.13976493537425994 - MAE: 0.2809897342394576 Validation loss : 0.14930716984801823 - MAE: 0.2915501156778208 Epoch: 82 Training loss: 0.13674694985151292 - MAE: 0.27840960478937093 Validation loss : 0.14946243580844668 - MAE: 0.2906570599625246 Epoch: 83 Training loss: 0.1379085114598274 - MAE: 0.27905460035509927 Validation loss : 0.14950388587183422 - MAE: 0.29112317700446616 Epoch: 84 Training loss: 0.14153614401817322 - MAE: 0.2835935983083774 Validation loss : 0.14871504406134287 - MAE: 0.2905583900515311 Epoch: 85 Training loss: 0.1404113781452179 - MAE: 0.2815505957403188 Validation loss : 0.15186349633667204 - MAE: 0.2935015096363652 Epoch: 86 Training loss: 0.1391930565237999 - MAE: 0.27938914945504073 Validation loss : 0.15015774303012425 - MAE: 0.29175260904666095 Epoch: 87 Training loss: 0.13965154111385344 - MAE: 0.2808129172644999 Validation loss : 0.14953040414386326 - MAE: 0.29217369917685887 Epoch: 88 Training loss: 0.14032937526702882 - MAE: 0.2820906583856728 Validation loss : 0.14954294512669244 - MAE: 0.2906296771253765 Epoch: 89 Training loss: 0.13956192165613174 - MAE: 0.279473451500942 Validation loss : 0.15065168423785102 - MAE: 0.29133002328368285 Epoch: 90 Training loss: 0.14047872692346572 - MAE: 0.2801611359909748 Validation loss : 0.14869611213604608 - MAE: 0.2914427528019158 Epoch: 91 Training loss: 0.14015074878931044 - MAE: 0.28182570327659834 Validation loss : 0.15071149749888313 - MAE: 0.2923330178798639 Epoch: 92 Training loss: 0.14030003398656846 - MAE: 0.2805739854352897 Validation loss : 0.1476295126809014 - MAE: 0.2897238746936595 Epoch: 93 Training loss: 0.1381376227736473 - MAE: 0.27943396254710007 Validation loss : 0.15042445477512148 - MAE: 0.2910839063768687 Epoch: 94 Training loss: 0.13906821191310884 - MAE: 0.27927564341730177 Validation loss : 0.1470845432745086 - MAE: 0.28945548179262404 Epoch: 95 Training loss: 0.13878683894872665 - MAE: 0.27950185756391205 Validation loss : 0.14802363597684437 - MAE: 0.2904005203523502 Epoch: 96 Training loss: 0.1400198295712471 - MAE: 0.28227933824874535 Validation loss : 0.14741118169493145 - MAE: 0.2886919057055488 Epoch: 97 Training loss: 0.14008406758308412 - MAE: 0.2820675306746204 Validation loss : 0.1483982594476806 - MAE: 0.2895298978938563 Epoch: 98 Training loss: 0.13766043066978453 - MAE: 0.27798967107381206 Validation loss : 0.1500189420249727 - MAE: 0.29136640929113017 Epoch: 99 Training loss: 0.14085446119308473 - MAE: 0.28138468637744457 Validation loss : 0.147224517332183 - MAE: 0.28914512721321917 Epoch: 100 Training loss: 0.140465350151062 - MAE: 0.28220289946670496 Validation loss : 0.14812367161115012 - MAE: 0.289046826605416 Epoch: 101 Training loss: 0.13990859061479569 - MAE: 0.2795642386535082 Validation loss : 0.1493135475450092 - MAE: 0.29040059011073405 Epoch: 102 Training loss: 0.14165273666381836 - MAE: 0.28245179865396797 Validation loss : 0.14880059079991448 - MAE: 0.29071234075534924 Epoch: 103 Training loss: 0.1388886308670044 - MAE: 0.279369814576424 Validation loss : 0.1476850998070505 - MAE: 0.2903485271171268 Epoch: 104 Training loss: 0.13909724473953247 - MAE: 0.28083273552929844 Validation loss : 0.14797810382313198 - MAE: 0.28941232666977995 Epoch: 105 Training loss: 0.1393595016002655 - MAE: 0.2783428221019483 Validation loss : 0.14831788175635868 - MAE: 0.2905713458472301 Epoch: 106 Training loss: 0.1390196332335472 - MAE: 0.27999429469540527 Validation loss : 0.1493500214484003 - MAE: 0.29255232062385816 Epoch: 107 Training loss: 0.1404520383477211 - MAE: 0.28146161503782474 Validation loss : 0.14878038648102018 - MAE: 0.2906487705342995 Epoch: 108 Training loss: 0.14004070401191712 - MAE: 0.28177979907550266 Validation loss : 0.14799664749039543 - MAE: 0.28897706210829416 Epoch: 109 Training loss: 0.1383381900191307 - MAE: 0.2784476944130519 Validation loss : 0.1490860970483886 - MAE: 0.2903748192836748 Epoch: 110 Training loss: 0.14010255575180053 - MAE: 0.2814791462622622 Validation loss : 0.14862897247076035 - MAE: 0.2912308111745315 Epoch: 111 Training loss: 0.1402219730615616 - MAE: 0.2815668657801355 Validation loss : 0.14855213463306427 - MAE: 0.2909928236013806 Epoch: 112 Training loss: 0.13920035630464553 - MAE: 0.281543453418757 Validation loss : 0.14755674368805355 - MAE: 0.2900560617682023 Epoch: 113 Training loss: 0.1400330564379692 - MAE: 0.2802442792672325 Validation loss : 0.14960995233721203 - MAE: 0.29297903682515697 Epoch: 114 Training loss: 0.13953691601753235 - MAE: 0.2798329911916557 Validation loss : 0.14831144445472294 - MAE: 0.2901465696869431 Epoch: 115 Training loss: 0.13693355649709701 - MAE: 0.2773904079655369 Validation loss : 0.14724988407558864 - MAE: 0.2901089629035545 Epoch: 116 Training loss: 0.13829326421022414 - MAE: 0.28039887425826104 Validation loss : 0.14879283474551308 - MAE: 0.29064504580363937 Epoch: 117 Training loss: 0.14090506106615067 - MAE: 0.2826000565784893 Validation loss : 0.1477228750785192 - MAE: 0.2903357764542001 Epoch: 118 Training loss: 0.1429896029829979 - MAE: 0.2845150458178373 Validation loss : 0.1520556633671125 - MAE: 0.2927788288379692 Epoch: 119 Training loss: 0.14179373949766158 - MAE: 0.28321784029712616 Validation loss : 0.1478977551062902 - MAE: 0.28899535825863276 Epoch: 120 Training loss: 0.13644426107406615 - MAE: 0.277860572559567 Validation loss : 0.1487949498825603 - MAE: 0.289971918515328 Epoch: 121 Training loss: 0.13941168040037155 - MAE: 0.28066886387000756 Validation loss : 0.14904049038887024 - MAE: 0.2892359740394001 Epoch: 122 Training loss: 0.13944829851388932 - MAE: 0.28015464875184926 Validation loss : 0.14853730301062265 - MAE: 0.28979281078938246 Epoch: 123 Training loss: 0.13739350140094758 - MAE: 0.27918069067817697 Validation loss : 0.14981382257408565 - MAE: 0.29127630466726984 Epoch: 124 Training loss: 0.13999282330274582 - MAE: 0.2798265373231701 Validation loss : 0.14869198782576454 - MAE: 0.29036692179389667 Epoch: 125 Training loss: 0.14108887642621995 - MAE: 0.28223551065679303 Validation loss : 0.1493358156747288 - MAE: 0.29129564350186565 Epoch: 126 Training loss: 0.13976523160934448 - MAE: 0.2817742714626282 Validation loss : 0.14911426686578327 - MAE: 0.29106398685631213 Epoch: 127 Training loss: 0.14017222493886947 - MAE: 0.28179540480169774 Validation loss : 0.14842878530422846 - MAE: 0.2901467176758952 Epoch: 128 Training loss: 0.1406565561890602 - MAE: 0.28163913944949737 Validation loss : 0.1479131720132298 - MAE: 0.28919518276952677 Epoch: 129 Training loss: 0.13942452102899552 - MAE: 0.2793156197663516 Validation loss : 0.1511403951379988 - MAE: 0.29067412905683 Epoch: 130 Training loss: 0.14031505227088928 - MAE: 0.2804210714866914 Validation loss : 0.14765288515223396 - MAE: 0.288500744825017 Epoch: 131 Training loss: 0.13649964869022368 - MAE: 0.2774726677017398 Validation loss : 0.14891337934467527 - MAE: 0.2914018373625627 Epoch: 132 Training loss: 0.1405644327402115 - MAE: 0.2804009316942137 Validation loss : 0.1483287529812919 - MAE: 0.2907619964928693 Epoch: 133 Training loss: 0.14118699222803116 - MAE: 0.2819625654924857 Validation loss : 0.15040278683106104 - MAE: 0.2910529233222426 Epoch: 134 Training loss: 0.14081039100885392 - MAE: 0.28238484691454196 Validation loss : 0.14994350406858656 - MAE: 0.2906808160616756 Epoch: 135 Training loss: 0.13862200796604157 - MAE: 0.27899754217163214 Validation loss : 0.14944083078040016 - MAE: 0.2906151099314951 Epoch: 136 Training loss: 0.13931725412607193 - MAE: 0.28030006822907155 Validation loss : 0.1495074381430944 - MAE: 0.29074428772408845 Epoch: 137 Training loss: 0.1403974002599716 - MAE: 0.2805740067883291 Validation loss : 0.148383229970932 - MAE: 0.29007782636865753 Epoch: 138 Training loss: 0.1385694068670273 - MAE: 0.27983454104510636 Validation loss : 0.1512098577287462 - MAE: 0.2919661208163271 Epoch: 139 Training loss: 0.1404304113984108 - MAE: 0.28148055927808185 Validation loss : 0.1489351209666994 - MAE: 0.29030302398224317 Epoch: 140 Training loss: 0.1390806904435158 - MAE: 0.27891810870400857 Validation loss : 0.1524661953250567 - MAE: 0.29273273639235975 Epoch: 141 Training loss: 0.13796700328588485 - MAE: 0.27966280744881894 Validation loss : 0.15020373629199135 - MAE: 0.2912314458069829 Epoch: 142 Training loss: 0.14071892708539963 - MAE: 0.2816539302440052 Validation loss : 0.15058360497156778 - MAE: 0.2914712531041945 Epoch: 143 Training loss: 0.14057636469602586 - MAE: 0.28242232736529516 Validation loss : 0.1506333847840627 - MAE: 0.29140815832807077 Epoch: 144 Training loss: 0.13925146371126174 - MAE: 0.2799103959499411 Validation loss : 0.14928915599981943 - MAE: 0.2897573463488531 Epoch: 145 Training loss: 0.1398252120614052 - MAE: 0.28164205446635043 Validation loss : 0.14851086835066477 - MAE: 0.29021297413113184 Epoch: 146 Training loss: 0.13817338794469833 - MAE: 0.27937391482438806 Validation loss : 0.1498860807882415 - MAE: 0.29185029747881874 Epoch: 147 Training loss: 0.14039424389600755 - MAE: 0.28106582531667956 Validation loss : 0.14853160911136204 - MAE: 0.29055607567586406 Epoch: 148 Training loss: 0.14145232886075973 - MAE: 0.2822118953799359 Validation loss : 0.15043112718396717 - MAE: 0.2923702620366525 Epoch: 149 Training loss: 0.14144912719726563 - MAE: 0.281877632162409 Validation loss : 0.1527728090683619 - MAE: 0.2934056300021226 Epoch: 150 Training loss: 0.14076999455690384 - MAE: 0.282006950149505 Validation loss : 0.15191659910811317 - MAE: 0.29212817923698803 Epoch: 151 Training loss: 0.14104728668928146 - MAE: 0.282059518005827 Validation loss : 0.15123689423004785 - MAE: 0.29217216756303294 Epoch: 152 Training loss: 0.14049483269453047 - MAE: 0.2816058638889221 Validation loss : 0.14788557837406793 - MAE: 0.29005179357861427 Epoch: 153 Training loss: 0.13666647017002106 - MAE: 0.276851346022365 Validation loss : 0.1481863541735543 - MAE: 0.2906076278179208 Epoch: 154 Training loss: 0.1380725646018982 - MAE: 0.27842498943152266 Validation loss : 0.14886879755391014 - MAE: 0.29055297838102373 Epoch: 155 Training loss: 0.13862104922533036 - MAE: 0.2793241343310853 Validation loss : 0.14891281558407676 - MAE: 0.29119474966859027 Epoch: 156 Training loss: 0.13700734436511994 - MAE: 0.278056366301915 Validation loss : 0.1506768903798527 - MAE: 0.2922085560624545 Epoch: 157 Training loss: 0.13953308165073394 - MAE: 0.281015862464819 Validation loss : 0.1483795220653216 - MAE: 0.29058860133117714 Epoch: 158 Training loss: 0.14014002352952956 - MAE: 0.28094309193212397 Validation loss : 0.1488533839583397 - MAE: 0.2904079557435194 Epoch: 159 Training loss: 0.1392805379629135 - MAE: 0.280401866801906 Validation loss : 0.152536627319124 - MAE: 0.29208377007193054 Epoch: 160 Training loss: 0.13968930572271346 - MAE: 0.2801028720194759 Validation loss : 0.15070715049902597 - MAE: 0.2915361243575389 Epoch: 161 Training loss: 0.1379876920580864 - MAE: 0.2779706198029852 Validation loss : 0.14963952700297037 - MAE: 0.29140363323705737 Epoch: 162 Training loss: 0.13882436752319335 - MAE: 0.2803142060072604 Validation loss : 0.14831043117576176 - MAE: 0.29090746702578535 Epoch: 163 Training loss: 0.14012926697731018 - MAE: 0.2826213295889931 Validation loss : 0.15037808484501308 - MAE: 0.2921038937559237 Epoch: 164 Training loss: 0.13816417813301085 - MAE: 0.27881172247852454 Validation loss : 0.14971058236228096 - MAE: 0.29022949264590414 Epoch: 165 Training loss: 0.14038401901721953 - MAE: 0.28114290350520105 Validation loss : 0.14926749053928587 - MAE: 0.2900267579172161 Epoch: 166 Training loss: 0.14115286707878114 - MAE: 0.2812040498053005 Validation loss : 0.15122641126314798 - MAE: 0.29175891756760863 Epoch: 167 Training loss: 0.13707010865211486 - MAE: 0.2777921438305016 Validation loss : 0.150983441207144 - MAE: 0.29142422296616527 Epoch: 168 Training loss: 0.13876796901226043 - MAE: 0.2802754449333229 Validation loss : 0.14823240455653933 - MAE: 0.2911944734775949 Epoch: 169 Training loss: 0.1400000700354576 - MAE: 0.27962448113148264 Validation loss : 0.15024596121576098 - MAE: 0.2927376945330595 Epoch: 170 Training loss: 0.13957252651453017 - MAE: 0.27999991303737254 Validation loss : 0.15100336488750246 - MAE: 0.29298232371488797 Epoch: 171 Training loss: 0.1379392111301422 - MAE: 0.28030924084273134 Validation loss : 0.1500499521692594 - MAE: 0.29190070752057323 Epoch: 172 Training loss: 0.13905995577573776 - MAE: 0.27959100752691673 Validation loss : 0.1501365229487419 - MAE: 0.291390080732848 Epoch: 173 Training loss: 0.13984983891248703 - MAE: 0.28052682384061434 Validation loss : 0.15079624785317314 - MAE: 0.2914017756619142 Epoch: 174 Training loss: 0.13995950669050217 - MAE: 0.28073402267442327 Validation loss : 0.15117037296295166 - MAE: 0.2913482358293418 Epoch: 175 Training loss: 0.13972256630659102 - MAE: 0.2810884447912753 Validation loss : 0.1483582788043552 - MAE: 0.2891574760822575 Epoch: 176 Training loss: 0.1396074891090393 - MAE: 0.2802075005571549 Validation loss : 0.15012793242931366 - MAE: 0.29115621681530746 Epoch: 177 Training loss: 0.14028172552585602 - MAE: 0.2809140783358514 Validation loss : 0.15086209111743504 - MAE: 0.2916531135350998 Epoch: 178 Training loss: 0.13791372925043105 - MAE: 0.27898197003406183 Validation loss : 0.15098679976330864 - MAE: 0.2925168924383861 Epoch: 179 Training loss: 0.1382387799024582 - MAE: 0.27801012015838344 Validation loss : 0.1535450733370251 - MAE: 0.29484870415135217 Epoch: 180 Training loss: 0.1407102045416832 - MAE: 0.28173970451231156 Validation loss : 0.1496893283393648 - MAE: 0.2922256774422862 Epoch: 181 Training loss: 0.13767244040966034 - MAE: 0.27825487061974347 Validation loss : 0.14896611786550945 - MAE: 0.2912224345710289 Epoch: 182 Training loss: 0.1394115325808525 - MAE: 0.28027065459265726 Validation loss : 0.14866195287969378 - MAE: 0.29134036191775803 Epoch: 183 Training loss: 0.13798033624887465 - MAE: 0.27942216029668876 Validation loss : 0.1484985484017266 - MAE: 0.2912727374193501 |