File size: 8,930 Bytes
a9614e2
1c266c2
573d184
138c452
085fc0e
a8e4a32
60030c6
43aca8f
bced6f4
3c0ba3e
05cb146
2755fdc
6d43361
9ba6b1c
bc4b6fa
fd66402
8ac5557
0a5b886
7582378
7943a4b
1d207e4
fc0b9fb
778b2ee
d8f234d
b8f30b4
07d6b01
b1e9821
5d63be6
c240e23
7fd6a30
753d906
7988c17
f9cc4b4
2563dd5
3b32d9d
78c220f
13dd3c9
3645ffc
a67ad3c
1791d1d
417295b
9d01042
4692ce2
ee702fe
9ad9b31
8a55390
af3dc92
255d975
b5c59b2
6291ba6
910096e
f1cb193
5ecc42f
f52fd05
4aa4325
5b2f689
0ca188f
90fba87
3b6adf1
d7b4484
b3d9166
822020b
5f1eb05
5a03441
36bc79e
ff030dd
927b3c1
64f46f0
749183e
2c6fb37
bff4a2d
46697fc
0413f7a
5ee6dd5
419fd28
eb700da
64c500b
d0056fd
66982b8
1d41cd7
1c356ba
d930327
fcc9b1f
1399e97
04ba03c
ed8b3f4
025040b
13b93d7
d26ef5a
a37fc5f
cf6faf1
86a6afb
40f3e9f
7cbbdc1
f5369b2
63b151c
eaadc50
6354a95
0d8ed18
8182faf
f9e751c
cb1f440
8a3fa68
21062cf
7ce2a2c
1f989a1
3d97e81
23665ab
57ada11
4a041ea
91a97e6
f4e0960
42a05eb
214e4e3
a4c8ea3
1d48055
55af590
864f0d5
d12921d
b7fd40a
e4b2e16
35063ea
c287647
9cc161b
16df9c1
8c239b2
e1ffb75
f067b6b
5707b55
06dbd2d
ba89237
748bfcb
ce6bb9d
a0c316e
0e41bee
a7c336c
284862e
e6bfda7
ff1b253
7f85e3c
39360cd
bf774fd
751781f
13f6f78
a14234f
2e1113b
26dbf57
48ecfcf
c37f189
5e1b947
76ec1ac
bc66fa8
71741a1
6fe9bdc
cba9e16
e6484a9
e9f9dc2
e2560d5
94d08de
ab2ab8b
306c777
b0daf82
7161dc2
ec8b412
6e7f355
1dbcf89
74c5bce
8f28059
cd5e1ba
50d54b4
e3854b6
47f5e22
0e66be8
0596de2
67a7610
eacd439
610aa7d
76212f8
5a43060
94586e7
6fae1c1
8ce35ef
de84ad0
71f0a6e
7988ea4
2ce4e82
0b0e4ff
e51065c
c85d65b
be9081c
d27e527
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
Started at: 21:34:05
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.32175946950912476 - MAE: 0.4437903503551826
Validation loss : 0.19621961481041378 - MAE: 0.3388283822410594
Epoch: 1
Training loss: 0.21072698533535003 - MAE: 0.35116530289867337
Validation loss : 0.1876767095592287 - MAE: 0.33628140968919695
Epoch: 2
Training loss: 0.1893942326307297 - MAE: 0.3312522516218666
Validation loss : 0.1748287214173211 - MAE: 0.3213797447248068
Epoch: 3
Training loss: 0.17969159841537474 - MAE: 0.3215382240664255
Validation loss : 0.1697687026527193 - MAE: 0.31446870224950263
Epoch: 4
Training loss: 0.17220519423484804 - MAE: 0.3129526796594085
Validation loss : 0.16811848680178323 - MAE: 0.3117262145282266
Epoch: 5
Training loss: 0.17296144247055054 - MAE: 0.31494873624825814
Validation loss : 0.16849000917540657 - MAE: 0.31104514179599985
Epoch: 6
Training loss: 0.16641035348176955 - MAE: 0.30768475494362546
Validation loss : 0.16421516074074638 - MAE: 0.30715220958263423
Epoch: 7
Training loss: 0.16057054400444032 - MAE: 0.30329494898002457
Validation loss : 0.161027698053254 - MAE: 0.304297376785587
Epoch: 8
Training loss: 0.1566170272231102 - MAE: 0.29957014870206655
Validation loss : 0.15982638630602095 - MAE: 0.30397747682822174
Epoch: 9
Training loss: 0.15673983812332154 - MAE: 0.3000540458404174
Validation loss : 0.159184659520785 - MAE: 0.3035235378542429
Epoch: 10
Training loss: 0.15630604147911073 - MAE: 0.29690365842430627
Validation loss : 0.15921704471111298 - MAE: 0.30240467396157955
Epoch: 11
Training loss: 0.15595020622015 - MAE: 0.29754135005638765
Validation loss : 0.15686986181471083 - MAE: 0.30119996351152656
Epoch: 12
Training loss: 0.15274528950452804 - MAE: 0.2944161972508913
Validation loss : 0.15652166141404045 - MAE: 0.3012713923501961
Epoch: 13
Training loss: 0.15205995708703995 - MAE: 0.29424324063629004
Validation loss : 0.1575678288936615 - MAE: 0.3004794443503115
Epoch: 14
Training loss: 0.15087180227041244 - MAE: 0.292504579327589
Validation loss : 0.15574459234873453 - MAE: 0.29932356111019714
Epoch: 15
Training loss: 0.1518820345401764 - MAE: 0.2940904971897683
Validation loss : 0.15581322544150883 - MAE: 0.30048684662514935
Epoch: 16
Training loss: 0.14916340589523316 - MAE: 0.2900369708605816
Validation loss : 0.15502946575482687 - MAE: 0.2986623058669649
Epoch: 17
Training loss: 0.14969733864068985 - MAE: 0.2916869417468108
Validation loss : 0.15631223718325296 - MAE: 0.2994015598567933
Epoch: 18
Training loss: 0.14621972769498826 - MAE: 0.2875086269286061
Validation loss : 0.1557358337773217 - MAE: 0.29931970436403404
Epoch: 19
Training loss: 0.14798570185899734 - MAE: 0.29024787778757843
Validation loss : 0.15424930387073094 - MAE: 0.2984811820958494
Epoch: 20
Training loss: 0.14769238144159316 - MAE: 0.2896109423923894
Validation loss : 0.1536422868569692 - MAE: 0.2987445844262027
Epoch: 21
Training loss: 0.14361368536949157 - MAE: 0.2842206176667335
Validation loss : 0.1533755792511834 - MAE: 0.29642898867488277
Epoch: 22
Training loss: 0.14456430345773696 - MAE: 0.2860957867005398
Validation loss : 0.1542035871081882 - MAE: 0.2971775973100257
Epoch: 23
Training loss: 0.1426533755660057 - MAE: 0.28512021628758083
Validation loss : 0.15245803362793392 - MAE: 0.2961228783995986
Epoch: 24
Training loss: 0.14542177826166153 - MAE: 0.2855342243309425
Validation loss : 0.15294104317824045 - MAE: 0.29643245601047447
Epoch: 25
Training loss: 0.14522234290838243 - MAE: 0.2851345876886818
Validation loss : 0.15342944694889915 - MAE: 0.2958033843008907
Epoch: 26
Training loss: 0.14386597275733948 - MAE: 0.2838243646140143
Validation loss : 0.1515038808186849 - MAE: 0.2936464496284943
Epoch: 27
Training loss: 0.1415349954366684 - MAE: 0.2810586437104536
Validation loss : 0.1507236527072059 - MAE: 0.29409404478735646
Epoch: 28
Training loss: 0.14115683376789093 - MAE: 0.2815221038197439
Validation loss : 0.15061336921321022 - MAE: 0.29280129784932213
Epoch: 29
Training loss: 0.14186541587114335 - MAE: 0.28239653007887255
Validation loss : 0.15151139597098032 - MAE: 0.2953867482779256
Epoch: 30
Training loss: 0.14276256740093232 - MAE: 0.283208699006944
Validation loss : 0.1504600097735723 - MAE: 0.29276670315056585
Epoch: 31
Training loss: 0.14241950929164887 - MAE: 0.28277882078383293
Validation loss : 0.1505332812666893 - MAE: 0.29293145394981956
Epoch: 32
Training loss: 0.14204519629478454 - MAE: 0.28192363607736504
Validation loss : 0.15044088496102226 - MAE: 0.2929076596067834
Epoch: 33
Training loss: 0.13920597046613692 - MAE: 0.2793298976614221
Validation loss : 0.15006180769867367 - MAE: 0.29292432343849906
Epoch: 34
Training loss: 0.1410813584923744 - MAE: 0.2815379053911258
Validation loss : 0.14930015967951882 - MAE: 0.2933547268525188
Epoch: 35
Training loss: 0.14150760889053346 - MAE: 0.2832031330097777
Validation loss : 0.15295125875208113 - MAE: 0.2936935655701593
Epoch: 36
Training loss: 0.1405733221769333 - MAE: 0.2821643593441243
Validation loss : 0.14867112785577774 - MAE: 0.2921335742814868
Epoch: 37
Training loss: 0.14206359952688216 - MAE: 0.28261241361934286
Validation loss : 0.14988169405195448 - MAE: 0.2927604772074144
Epoch: 38
Training loss: 0.14064243495464324 - MAE: 0.28185439579063915
Validation loss : 0.1515509072277281 - MAE: 0.29319069832718325
Epoch: 39
Training loss: 0.14297345608472825 - MAE: 0.282847817609895
Validation loss : 0.15042786465750801 - MAE: 0.29282722784595605
Epoch: 40
Training loss: 0.13990240722894667 - MAE: 0.27964537481988466
Validation loss : 0.14972211172183356 - MAE: 0.29260355981671565
Epoch: 41
Training loss: 0.13921964198350906 - MAE: 0.28078253547199195
Validation loss : 0.14965214083592096 - MAE: 0.2924700199853063
Epoch: 42
Training loss: 0.1393338233232498 - MAE: 0.28024463401390987
Validation loss : 0.1507701873779297 - MAE: 0.29355734742186085
Epoch: 43
Training loss: 0.1382669734954834 - MAE: 0.27992590956331337
Validation loss : 0.14967524343066746 - MAE: 0.29149855699855537
Epoch: 44
Training loss: 0.14029370576143266 - MAE: 0.28265328374425863
Validation loss : 0.1491319098406368 - MAE: 0.29219012845466735
Epoch: 45
Training loss: 0.14041882872581482 - MAE: 0.2809566443072132
Validation loss : 0.14969545520014232 - MAE: 0.2909081862153574
Epoch: 46
Training loss: 0.140560123026371 - MAE: 0.28191663739835304
Validation loss : 0.14977325167920855 - MAE: 0.291604275381065
Epoch: 47
Training loss: 0.1414617270231247 - MAE: 0.28282842532543745
Validation loss : 0.1506698860062493 - MAE: 0.2930552446484735
Epoch: 48
Training loss: 0.1373920688033104 - MAE: 0.27917131881687096
Validation loss : 0.15224417712953356 - MAE: 0.2934837599479946
Epoch: 49
Training loss: 0.14192317247390748 - MAE: 0.2842585649178634
Validation loss : 0.14885787086354363 - MAE: 0.29192342005218896
Epoch: 50
Training loss: 0.14062062561511993 - MAE: 0.2808979887022064
Validation loss : 0.15054219878382152 - MAE: 0.2924492759311905
Epoch: 51
Training loss: 0.13990722686052323 - MAE: 0.2817114196110173
Validation loss : 0.1506080221798685 - MAE: 0.29247582518131027
Epoch: 52
Training loss: 0.1394276064634323 - MAE: 0.2798831110400087
Validation loss : 0.1490704102648629 - MAE: 0.291844280782309
Epoch: 53
Training loss: 0.14104595750570298 - MAE: 0.28180114573645876
Validation loss : 0.14941397143734825 - MAE: 0.2933880270519655
Epoch: 54
Training loss: 0.139021515250206 - MAE: 0.2793780218867486
Validation loss : 0.1490028988983896 - MAE: 0.29192548589513484
Epoch: 55
Training loss: 0.1409096112847328 - MAE: 0.2816018781967446
Validation loss : 0.1494157827562756 - MAE: 0.2919783378546487
Epoch: 56
Training loss: 0.14150582283735275 - MAE: 0.28397171057514964
Validation loss : 0.1507257032725546 - MAE: 0.29278463527362475
Epoch: 57
Training loss: 0.1412457060813904 - MAE: 0.28194929681996234
Validation loss : 0.1500752924217118 - MAE: 0.291890501089955
Epoch: 58
Training loss: 0.14060731261968612 - MAE: 0.2826944446609872
Validation loss : 0.15134404020176995 - MAE: 0.2921265934844939
Epoch: 59
Training loss: 0.13857035160064698 - MAE: 0.27917573503863885
Validation loss : 0.1502113996280564 - MAE: 0.2916726146711328
Epoch: 60
Training loss: 0.14019066840410233 - MAE: 0.2816340620891049
Validation loss : 0.14821443955103555 - MAE: 0.29117830198396194
Epoch: 61
Training loss: 0.1397022610902786 - MAE: 0.28089572600032964
Validation loss : 0.1513935931854778 - MAE: 0.2917060470497501
Epoch: 62
Training loss: 0.14070956617593766 - MAE: 0.28241979367979564
Validation loss : 0.1501304821835624 - MAE: 0.2912750997306758