File size: 18,553 Bytes
a9614e2
1c266c2
573d184
138c452
085fc0e
a8e4a32
60030c6
43aca8f
bced6f4
3c0ba3e
05cb146
2755fdc
6d43361
9ba6b1c
bc4b6fa
fd66402
8ac5557
0a5b886
7582378
7943a4b
1d207e4
fc0b9fb
778b2ee
d8f234d
b8f30b4
07d6b01
b1e9821
5d63be6
c240e23
7fd6a30
753d906
7988c17
f9cc4b4
2563dd5
3b32d9d
78c220f
13dd3c9
3645ffc
a67ad3c
1791d1d
417295b
9d01042
4692ce2
ee702fe
9ad9b31
8a55390
af3dc92
255d975
b5c59b2
6291ba6
910096e
f1cb193
5ecc42f
f52fd05
4aa4325
5b2f689
0ca188f
90fba87
3b6adf1
d7b4484
b3d9166
822020b
5f1eb05
5a03441
36bc79e
ff030dd
927b3c1
64f46f0
749183e
2c6fb37
bff4a2d
46697fc
0413f7a
5ee6dd5
419fd28
eb700da
64c500b
d0056fd
66982b8
1d41cd7
1c356ba
d930327
fcc9b1f
1399e97
04ba03c
ed8b3f4
025040b
13b93d7
d26ef5a
a37fc5f
cf6faf1
86a6afb
40f3e9f
7cbbdc1
f5369b2
63b151c
eaadc50
6354a95
0d8ed18
8182faf
f9e751c
cb1f440
8a3fa68
21062cf
7ce2a2c
1f989a1
3d97e81
23665ab
57ada11
4a041ea
91a97e6
f4e0960
42a05eb
214e4e3
a4c8ea3
1d48055
55af590
864f0d5
d12921d
b7fd40a
e4b2e16
35063ea
c287647
9cc161b
16df9c1
8c239b2
e1ffb75
f067b6b
5707b55
06dbd2d
ba89237
748bfcb
ce6bb9d
a0c316e
0e41bee
a7c336c
284862e
e6bfda7
ff1b253
7f85e3c
39360cd
bf774fd
751781f
13f6f78
a14234f
2e1113b
26dbf57
48ecfcf
c37f189
5e1b947
76ec1ac
bc66fa8
71741a1
6fe9bdc
cba9e16
e6484a9
e9f9dc2
e2560d5
94d08de
ab2ab8b
306c777
b0daf82
7161dc2
ec8b412
6e7f355
1dbcf89
74c5bce
8f28059
cd5e1ba
50d54b4
e3854b6
47f5e22
0e66be8
0596de2
67a7610
eacd439
610aa7d
76212f8
5a43060
94586e7
6fae1c1
8ce35ef
de84ad0
71f0a6e
7988ea4
2ce4e82
0b0e4ff
e51065c
c85d65b
be9081c
d27e527
4fe4ad7
edff910
8f89e59
f773786
0c18394
e5713c6
4257128
587cfee
2855675
c951d8f
26d82b2
668d135
cf661ef
5d6ba54
e07588e
0bf1606
7ba5a4f
5ccf23e
04c8203
3ce5140
0142e2d
aa8c5f1
14e679e
5e114da
3fa5453
2d659e3
d996cf0
56ae8c1
0ff9e6e
cef854e
db07dce
b013ee1
af9d814
c9d4c79
270f1db
0a8d8f5
757246c
b4908df
fb70983
1254ee4
9bd0527
adef98d
7fdaae7
f5aec4e
63e1b2c
d1c2ef2
cdfafca
d98eaa9
d8b886a
411270c
8f14fa0
0b4e95a
5fd9305
1c6a829
5359fa5
7e68e76
a3d1fa4
1d756fc
9789da4
b510afb
31e8acb
6ffc0ea
f558ed5
7989194
28131e9
55d92e6
5570ada
c30434c
b7bf10c
a50cfb9
328ddd3
ca5360b
df0a334
b9f3062
c20973c
3b2d473
6279f3b
2192d9b
6c052e1
35a62f3
9cd94e6
cea3250
1b7ab72
8c30aff
ba362ed
b20f700
dd5c50f
4eb1ab2
9935079
5b1f10c
fd34687
a095ee0
7117070
f93d704
222ad96
82126f9
bb5e24a
e96b271
7fcc0fd
36b0b74
ea78722
9584d30
72d3b07
de2672e
7faba00
e0bb648
10e4a00
0d74544
448c740
19de6ba
3f0b83f
14fcabb
b5a78e6
fcb1ba2
bc901e2
0c1ecf9
21aa32a
6650600
1881e39
186a8b7
146d0b6
47ff58e
3fab5ae
3c7dcd2
c59732a
c83ce85
db31097
a70bbcd
a365d1d
a53a7e5
e66e540
915d84c
3175739
0703d6e
96dd1ef
e9e6722
3056e8f
e51210c
ddfed9b
fc48cdd
84ddc15
99b7715
f63c516
a9a3095
5676aaa
134be0e
c73423c
2d505db
f94c81c
84df553
75845e8
82ba38b
171c0bb
8aee9c4
ec41b2e
5ee5721
5330555
a2e9cde
37a56dd
e3f480a
f1adf41
fd7181d
02a622c
cb5dfef
6607028
eb07595
9109f82
fc45727
fcdf3fb
0bba515
c1a350b
00bec49
8415958
c9e290a
0185638
e8b06c1
90b2b0d
486ccb8
6cffbbe
0e1162e
78a40bb
8f538a2
040d1b9
bd34f15
4377c2a
856a1c9
5320034
93e268a
9766d65
db83950
497cace
d0badb0
a536fb2
396420c
c61131b
abb5dfd
60c0337
687f13d
9b105ab
5f9a66b
1c0ed1e
e0896ef
f6401d3
809478d
5351220
198c64a
5774464
f87e23c
1f0928c
d8fdcc1
f0a5160
2095131
709aa53
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
Started at: 21:34:05
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.32175946950912476 - MAE: 0.4437903503551826
Validation loss : 0.19621961481041378 - MAE: 0.3388283822410594
Epoch: 1
Training loss: 0.21072698533535003 - MAE: 0.35116530289867337
Validation loss : 0.1876767095592287 - MAE: 0.33628140968919695
Epoch: 2
Training loss: 0.1893942326307297 - MAE: 0.3312522516218666
Validation loss : 0.1748287214173211 - MAE: 0.3213797447248068
Epoch: 3
Training loss: 0.17969159841537474 - MAE: 0.3215382240664255
Validation loss : 0.1697687026527193 - MAE: 0.31446870224950263
Epoch: 4
Training loss: 0.17220519423484804 - MAE: 0.3129526796594085
Validation loss : 0.16811848680178323 - MAE: 0.3117262145282266
Epoch: 5
Training loss: 0.17296144247055054 - MAE: 0.31494873624825814
Validation loss : 0.16849000917540657 - MAE: 0.31104514179599985
Epoch: 6
Training loss: 0.16641035348176955 - MAE: 0.30768475494362546
Validation loss : 0.16421516074074638 - MAE: 0.30715220958263423
Epoch: 7
Training loss: 0.16057054400444032 - MAE: 0.30329494898002457
Validation loss : 0.161027698053254 - MAE: 0.304297376785587
Epoch: 8
Training loss: 0.1566170272231102 - MAE: 0.29957014870206655
Validation loss : 0.15982638630602095 - MAE: 0.30397747682822174
Epoch: 9
Training loss: 0.15673983812332154 - MAE: 0.3000540458404174
Validation loss : 0.159184659520785 - MAE: 0.3035235378542429
Epoch: 10
Training loss: 0.15630604147911073 - MAE: 0.29690365842430627
Validation loss : 0.15921704471111298 - MAE: 0.30240467396157955
Epoch: 11
Training loss: 0.15595020622015 - MAE: 0.29754135005638765
Validation loss : 0.15686986181471083 - MAE: 0.30119996351152656
Epoch: 12
Training loss: 0.15274528950452804 - MAE: 0.2944161972508913
Validation loss : 0.15652166141404045 - MAE: 0.3012713923501961
Epoch: 13
Training loss: 0.15205995708703995 - MAE: 0.29424324063629004
Validation loss : 0.1575678288936615 - MAE: 0.3004794443503115
Epoch: 14
Training loss: 0.15087180227041244 - MAE: 0.292504579327589
Validation loss : 0.15574459234873453 - MAE: 0.29932356111019714
Epoch: 15
Training loss: 0.1518820345401764 - MAE: 0.2940904971897683
Validation loss : 0.15581322544150883 - MAE: 0.30048684662514935
Epoch: 16
Training loss: 0.14916340589523316 - MAE: 0.2900369708605816
Validation loss : 0.15502946575482687 - MAE: 0.2986623058669649
Epoch: 17
Training loss: 0.14969733864068985 - MAE: 0.2916869417468108
Validation loss : 0.15631223718325296 - MAE: 0.2994015598567933
Epoch: 18
Training loss: 0.14621972769498826 - MAE: 0.2875086269286061
Validation loss : 0.1557358337773217 - MAE: 0.29931970436403404
Epoch: 19
Training loss: 0.14798570185899734 - MAE: 0.29024787778757843
Validation loss : 0.15424930387073094 - MAE: 0.2984811820958494
Epoch: 20
Training loss: 0.14769238144159316 - MAE: 0.2896109423923894
Validation loss : 0.1536422868569692 - MAE: 0.2987445844262027
Epoch: 21
Training loss: 0.14361368536949157 - MAE: 0.2842206176667335
Validation loss : 0.1533755792511834 - MAE: 0.29642898867488277
Epoch: 22
Training loss: 0.14456430345773696 - MAE: 0.2860957867005398
Validation loss : 0.1542035871081882 - MAE: 0.2971775973100257
Epoch: 23
Training loss: 0.1426533755660057 - MAE: 0.28512021628758083
Validation loss : 0.15245803362793392 - MAE: 0.2961228783995986
Epoch: 24
Training loss: 0.14542177826166153 - MAE: 0.2855342243309425
Validation loss : 0.15294104317824045 - MAE: 0.29643245601047447
Epoch: 25
Training loss: 0.14522234290838243 - MAE: 0.2851345876886818
Validation loss : 0.15342944694889915 - MAE: 0.2958033843008907
Epoch: 26
Training loss: 0.14386597275733948 - MAE: 0.2838243646140143
Validation loss : 0.1515038808186849 - MAE: 0.2936464496284943
Epoch: 27
Training loss: 0.1415349954366684 - MAE: 0.2810586437104536
Validation loss : 0.1507236527072059 - MAE: 0.29409404478735646
Epoch: 28
Training loss: 0.14115683376789093 - MAE: 0.2815221038197439
Validation loss : 0.15061336921321022 - MAE: 0.29280129784932213
Epoch: 29
Training loss: 0.14186541587114335 - MAE: 0.28239653007887255
Validation loss : 0.15151139597098032 - MAE: 0.2953867482779256
Epoch: 30
Training loss: 0.14276256740093232 - MAE: 0.283208699006944
Validation loss : 0.1504600097735723 - MAE: 0.29276670315056585
Epoch: 31
Training loss: 0.14241950929164887 - MAE: 0.28277882078383293
Validation loss : 0.1505332812666893 - MAE: 0.29293145394981956
Epoch: 32
Training loss: 0.14204519629478454 - MAE: 0.28192363607736504
Validation loss : 0.15044088496102226 - MAE: 0.2929076596067834
Epoch: 33
Training loss: 0.13920597046613692 - MAE: 0.2793298976614221
Validation loss : 0.15006180769867367 - MAE: 0.29292432343849906
Epoch: 34
Training loss: 0.1410813584923744 - MAE: 0.2815379053911258
Validation loss : 0.14930015967951882 - MAE: 0.2933547268525188
Epoch: 35
Training loss: 0.14150760889053346 - MAE: 0.2832031330097777
Validation loss : 0.15295125875208113 - MAE: 0.2936935655701593
Epoch: 36
Training loss: 0.1405733221769333 - MAE: 0.2821643593441243
Validation loss : 0.14867112785577774 - MAE: 0.2921335742814868
Epoch: 37
Training loss: 0.14206359952688216 - MAE: 0.28261241361934286
Validation loss : 0.14988169405195448 - MAE: 0.2927604772074144
Epoch: 38
Training loss: 0.14064243495464324 - MAE: 0.28185439579063915
Validation loss : 0.1515509072277281 - MAE: 0.29319069832718325
Epoch: 39
Training loss: 0.14297345608472825 - MAE: 0.282847817609895
Validation loss : 0.15042786465750801 - MAE: 0.29282722784595605
Epoch: 40
Training loss: 0.13990240722894667 - MAE: 0.27964537481988466
Validation loss : 0.14972211172183356 - MAE: 0.29260355981671565
Epoch: 41
Training loss: 0.13921964198350906 - MAE: 0.28078253547199195
Validation loss : 0.14965214083592096 - MAE: 0.2924700199853063
Epoch: 42
Training loss: 0.1393338233232498 - MAE: 0.28024463401390987
Validation loss : 0.1507701873779297 - MAE: 0.29355734742186085
Epoch: 43
Training loss: 0.1382669734954834 - MAE: 0.27992590956331337
Validation loss : 0.14967524343066746 - MAE: 0.29149855699855537
Epoch: 44
Training loss: 0.14029370576143266 - MAE: 0.28265328374425863
Validation loss : 0.1491319098406368 - MAE: 0.29219012845466735
Epoch: 45
Training loss: 0.14041882872581482 - MAE: 0.2809566443072132
Validation loss : 0.14969545520014232 - MAE: 0.2909081862153574
Epoch: 46
Training loss: 0.140560123026371 - MAE: 0.28191663739835304
Validation loss : 0.14977325167920855 - MAE: 0.291604275381065
Epoch: 47
Training loss: 0.1414617270231247 - MAE: 0.28282842532543745
Validation loss : 0.1506698860062493 - MAE: 0.2930552446484735
Epoch: 48
Training loss: 0.1373920688033104 - MAE: 0.27917131881687096
Validation loss : 0.15224417712953356 - MAE: 0.2934837599479946
Epoch: 49
Training loss: 0.14192317247390748 - MAE: 0.2842585649178634
Validation loss : 0.14885787086354363 - MAE: 0.29192342005218896
Epoch: 50
Training loss: 0.14062062561511993 - MAE: 0.2808979887022064
Validation loss : 0.15054219878382152 - MAE: 0.2924492759311905
Epoch: 51
Training loss: 0.13990722686052323 - MAE: 0.2817114196110173
Validation loss : 0.1506080221798685 - MAE: 0.29247582518131027
Epoch: 52
Training loss: 0.1394276064634323 - MAE: 0.2798831110400087
Validation loss : 0.1490704102648629 - MAE: 0.291844280782309
Epoch: 53
Training loss: 0.14104595750570298 - MAE: 0.28180114573645876
Validation loss : 0.14941397143734825 - MAE: 0.2933880270519655
Epoch: 54
Training loss: 0.139021515250206 - MAE: 0.2793780218867486
Validation loss : 0.1490028988983896 - MAE: 0.29192548589513484
Epoch: 55
Training loss: 0.1409096112847328 - MAE: 0.2816018781967446
Validation loss : 0.1494157827562756 - MAE: 0.2919783378546487
Epoch: 56
Training loss: 0.14150582283735275 - MAE: 0.28397171057514964
Validation loss : 0.1507257032725546 - MAE: 0.29278463527362475
Epoch: 57
Training loss: 0.1412457060813904 - MAE: 0.28194929681996234
Validation loss : 0.1500752924217118 - MAE: 0.291890501089955
Epoch: 58
Training loss: 0.14060731261968612 - MAE: 0.2826944446609872
Validation loss : 0.15134404020176995 - MAE: 0.2921265934844939
Epoch: 59
Training loss: 0.13857035160064698 - MAE: 0.27917573503863885
Validation loss : 0.1502113996280564 - MAE: 0.2916726146711328
Epoch: 60
Training loss: 0.14019066840410233 - MAE: 0.2816340620891049
Validation loss : 0.14821443955103555 - MAE: 0.29117830198396194
Epoch: 61
Training loss: 0.1397022610902786 - MAE: 0.28089572600032964
Validation loss : 0.1513935931854778 - MAE: 0.2917060470497501
Epoch: 62
Training loss: 0.14070956617593766 - MAE: 0.28241979367979564
Validation loss : 0.1501304821835624 - MAE: 0.2912750997306758
Epoch: 63
Training loss: 0.1420893755555153 - MAE: 0.28372013668513324
Validation loss : 0.15038747092088064 - MAE: 0.2920179987433622
Epoch: 64
Training loss: 0.1394691723585129 - MAE: 0.2797968922342149
Validation loss : 0.1498954254719946 - MAE: 0.2930247953251004
Epoch: 65
Training loss: 0.14013640880584716 - MAE: 0.28057332071413116
Validation loss : 0.14924369752407074 - MAE: 0.29117805094765137
Epoch: 66
Training loss: 0.13910346567630769 - MAE: 0.27976148521760813
Validation loss : 0.15365933709674412 - MAE: 0.2938181280395696
Epoch: 67
Training loss: 0.14027394950389863 - MAE: 0.2824978199508478
Validation loss : 0.15143409785297182 - MAE: 0.2915787634959989
Epoch: 68
Training loss: 0.13879832834005357 - MAE: 0.2803663281034843
Validation loss : 0.14806738330258262 - MAE: 0.2899341093266117
Epoch: 69
Training loss: 0.14169045120477677 - MAE: 0.2840059654564348
Validation loss : 0.14932368530167472 - MAE: 0.2908179996636204
Epoch: 70
Training loss: 0.14286602079868316 - MAE: 0.28393754014878414
Validation loss : 0.15445275604724884 - MAE: 0.29465461309717
Epoch: 71
Training loss: 0.1397111463546753 - MAE: 0.2815673786321848
Validation loss : 0.14909260223309198 - MAE: 0.29124006221654514
Epoch: 72
Training loss: 0.13914978861808777 - MAE: 0.27919056115617275
Validation loss : 0.1496173938115438 - MAE: 0.29145609362558494
Epoch: 73
Training loss: 0.13840563476085663 - MAE: 0.2791765280391529
Validation loss : 0.15138710704114702 - MAE: 0.2922445451728484
Epoch: 74
Training loss: 0.13956822723150253 - MAE: 0.2806497885355469
Validation loss : 0.14996923257907233 - MAE: 0.2913337573766691
Epoch: 75
Training loss: 0.13771951258182524 - MAE: 0.2787076577823987
Validation loss : 0.1503291262520684 - MAE: 0.292124609001339
Epoch: 76
Training loss: 0.13808781534433365 - MAE: 0.27866726046665985
Validation loss : 0.14987454232242373 - MAE: 0.29144555209570666
Epoch: 77
Training loss: 0.1389396995306015 - MAE: 0.2791280405088736
Validation loss : 0.1490834727883339 - MAE: 0.29122889859275275
Epoch: 78
Training loss: 0.13928475528955458 - MAE: 0.28034273788540925
Validation loss : 0.14989753978119957 - MAE: 0.2910398033999464
Epoch: 79
Training loss: 0.13676762133836745 - MAE: 0.27751891334295836
Validation loss : 0.15044062005148995 - MAE: 0.29325611713326233
Epoch: 80
Training loss: 0.13998845636844634 - MAE: 0.28041881321085516
Validation loss : 0.14946704192294014 - MAE: 0.293321332550951
Epoch: 81
Training loss: 0.13976493537425994 - MAE: 0.2809897342394576
Validation loss : 0.14930716984801823 - MAE: 0.2915501156778208
Epoch: 82
Training loss: 0.13674694985151292 - MAE: 0.27840960478937093
Validation loss : 0.14946243580844668 - MAE: 0.2906570599625246
Epoch: 83
Training loss: 0.1379085114598274 - MAE: 0.27905460035509927
Validation loss : 0.14950388587183422 - MAE: 0.29112317700446616
Epoch: 84
Training loss: 0.14153614401817322 - MAE: 0.2835935983083774
Validation loss : 0.14871504406134287 - MAE: 0.2905583900515311
Epoch: 85
Training loss: 0.1404113781452179 - MAE: 0.2815505957403188
Validation loss : 0.15186349633667204 - MAE: 0.2935015096363652
Epoch: 86
Training loss: 0.1391930565237999 - MAE: 0.27938914945504073
Validation loss : 0.15015774303012425 - MAE: 0.29175260904666095
Epoch: 87
Training loss: 0.13965154111385344 - MAE: 0.2808129172644999
Validation loss : 0.14953040414386326 - MAE: 0.29217369917685887
Epoch: 88
Training loss: 0.14032937526702882 - MAE: 0.2820906583856728
Validation loss : 0.14954294512669244 - MAE: 0.2906296771253765
Epoch: 89
Training loss: 0.13956192165613174 - MAE: 0.279473451500942
Validation loss : 0.15065168423785102 - MAE: 0.29133002328368285
Epoch: 90
Training loss: 0.14047872692346572 - MAE: 0.2801611359909748
Validation loss : 0.14869611213604608 - MAE: 0.2914427528019158
Epoch: 91
Training loss: 0.14015074878931044 - MAE: 0.28182570327659834
Validation loss : 0.15071149749888313 - MAE: 0.2923330178798639
Epoch: 92
Training loss: 0.14030003398656846 - MAE: 0.2805739854352897
Validation loss : 0.1476295126809014 - MAE: 0.2897238746936595
Epoch: 93
Training loss: 0.1381376227736473 - MAE: 0.27943396254710007
Validation loss : 0.15042445477512148 - MAE: 0.2910839063768687
Epoch: 94
Training loss: 0.13906821191310884 - MAE: 0.27927564341730177
Validation loss : 0.1470845432745086 - MAE: 0.28945548179262404
Epoch: 95
Training loss: 0.13878683894872665 - MAE: 0.27950185756391205
Validation loss : 0.14802363597684437 - MAE: 0.2904005203523502
Epoch: 96
Training loss: 0.1400198295712471 - MAE: 0.28227933824874535
Validation loss : 0.14741118169493145 - MAE: 0.2886919057055488
Epoch: 97
Training loss: 0.14008406758308412 - MAE: 0.2820675306746204
Validation loss : 0.1483982594476806 - MAE: 0.2895298978938563
Epoch: 98
Training loss: 0.13766043066978453 - MAE: 0.27798967107381206
Validation loss : 0.1500189420249727 - MAE: 0.29136640929113017
Epoch: 99
Training loss: 0.14085446119308473 - MAE: 0.28138468637744457
Validation loss : 0.147224517332183 - MAE: 0.28914512721321917
Epoch: 100
Training loss: 0.140465350151062 - MAE: 0.28220289946670496
Validation loss : 0.14812367161115012 - MAE: 0.289046826605416
Epoch: 101
Training loss: 0.13990859061479569 - MAE: 0.2795642386535082
Validation loss : 0.1493135475450092 - MAE: 0.29040059011073405
Epoch: 102
Training loss: 0.14165273666381836 - MAE: 0.28245179865396797
Validation loss : 0.14880059079991448 - MAE: 0.29071234075534924
Epoch: 103
Training loss: 0.1388886308670044 - MAE: 0.279369814576424
Validation loss : 0.1476850998070505 - MAE: 0.2903485271171268
Epoch: 104
Training loss: 0.13909724473953247 - MAE: 0.28083273552929844
Validation loss : 0.14797810382313198 - MAE: 0.28941232666977995
Epoch: 105
Training loss: 0.1393595016002655 - MAE: 0.2783428221019483
Validation loss : 0.14831788175635868 - MAE: 0.2905713458472301
Epoch: 106
Training loss: 0.1390196332335472 - MAE: 0.27999429469540527
Validation loss : 0.1493500214484003 - MAE: 0.29255232062385816
Epoch: 107
Training loss: 0.1404520383477211 - MAE: 0.28146161503782474
Validation loss : 0.14878038648102018 - MAE: 0.2906487705342995
Epoch: 108
Training loss: 0.14004070401191712 - MAE: 0.28177979907550266
Validation loss : 0.14799664749039543 - MAE: 0.28897706210829416
Epoch: 109
Training loss: 0.1383381900191307 - MAE: 0.2784476944130519
Validation loss : 0.1490860970483886 - MAE: 0.2903748192836748
Epoch: 110
Training loss: 0.14010255575180053 - MAE: 0.2814791462622622
Validation loss : 0.14862897247076035 - MAE: 0.2912308111745315
Epoch: 111
Training loss: 0.1402219730615616 - MAE: 0.2815668657801355
Validation loss : 0.14855213463306427 - MAE: 0.2909928236013806
Epoch: 112
Training loss: 0.13920035630464553 - MAE: 0.281543453418757
Validation loss : 0.14755674368805355 - MAE: 0.2900560617682023
Epoch: 113
Training loss: 0.1400330564379692 - MAE: 0.2802442792672325
Validation loss : 0.14960995233721203 - MAE: 0.29297903682515697
Epoch: 114
Training loss: 0.13953691601753235 - MAE: 0.2798329911916557
Validation loss : 0.14831144445472294 - MAE: 0.2901465696869431
Epoch: 115
Training loss: 0.13693355649709701 - MAE: 0.2773904079655369
Validation loss : 0.14724988407558864 - MAE: 0.2901089629035545
Epoch: 116
Training loss: 0.13829326421022414 - MAE: 0.28039887425826104
Validation loss : 0.14879283474551308 - MAE: 0.29064504580363937
Epoch: 117
Training loss: 0.14090506106615067 - MAE: 0.2826000565784893
Validation loss : 0.1477228750785192 - MAE: 0.2903357764542001
Epoch: 118
Training loss: 0.1429896029829979 - MAE: 0.2845150458178373
Validation loss : 0.1520556633671125 - MAE: 0.2927788288379692
Epoch: 119
Training loss: 0.14179373949766158 - MAE: 0.28321784029712616
Validation loss : 0.1478977551062902 - MAE: 0.28899535825863276
Epoch: 120
Training loss: 0.13644426107406615 - MAE: 0.277860572559567
Validation loss : 0.1487949498825603 - MAE: 0.289971918515328
Epoch: 121
Training loss: 0.13941168040037155 - MAE: 0.28066886387000756
Validation loss : 0.14904049038887024 - MAE: 0.2892359740394001
Epoch: 122
Training loss: 0.13944829851388932 - MAE: 0.28015464875184926
Validation loss : 0.14853730301062265 - MAE: 0.28979281078938246
Epoch: 123
Training loss: 0.13739350140094758 - MAE: 0.27918069067817697
Validation loss : 0.14981382257408565 - MAE: 0.29127630466726984
Epoch: 124
Training loss: 0.13999282330274582 - MAE: 0.2798265373231701
Validation loss : 0.14869198782576454 - MAE: 0.29036692179389667
Epoch: 125
Training loss: 0.14108887642621995 - MAE: 0.28223551065679303
Validation loss : 0.1493358156747288 - MAE: 0.29129564350186565
Epoch: 126
Training loss: 0.13976523160934448 - MAE: 0.2817742714626282
Validation loss : 0.14911426686578327 - MAE: 0.29106398685631213
Epoch: 127
Training loss: 0.14017222493886947 - MAE: 0.28179540480169774
Validation loss : 0.14842878530422846 - MAE: 0.2901467176758952
Epoch: 128
Training loss: 0.1406565561890602 - MAE: 0.28163913944949737
Validation loss : 0.1479131720132298 - MAE: 0.28919518276952677
Epoch: 129
Training loss: 0.13942452102899552 - MAE: 0.2793156197663516
Validation loss : 0.1511403951379988 - MAE: 0.29067412905683
Epoch: 130
Training loss: 0.14031505227088928 - MAE: 0.2804210714866914
Validation loss : 0.14765288515223396 - MAE: 0.288500744825017
Epoch: 131
Training loss: 0.13649964869022368 - MAE: 0.2774726677017398
Validation loss : 0.14891337934467527 - MAE: 0.2914018373625627
Epoch: 132
Training loss: 0.1405644327402115 - MAE: 0.2804009316942137
Validation loss : 0.1483287529812919 - MAE: 0.2907619964928693
Epoch: 133
Training loss: 0.14118699222803116 - MAE: 0.2819625654924857
Validation loss : 0.15040278683106104 - MAE: 0.2910529233222426