File size: 11,242 Bytes
ee53a4a
50f1745
8a33155
79fb205
22be4a5
3f1ff0e
bbed904
14c983c
bb23377
0ea8477
8597b5f
a529489
8f5752b
d444cf9
2dfa794
782f4ba
2665515
8c5051b
84a2460
36464be
b4eab8a
b4c7480
bd5a96e
c248496
a3f3665
3572268
8f95e8a
0e3f82d
bd48add
1da9fd2
3759f70
e09bf1b
ec4c3b8
b891788
9c846de
c36a8be
3f890ba
aaf9dcb
9442814
64e656f
47314ea
b48fff4
c931c86
ec8e58d
28782de
d61067e
2cde297
dce64ed
1a44567
a894d88
f8b0b44
63524f2
3dc095f
915dd82
a5a47b9
15b72bc
5ba086a
6a406c0
fd9bd7c
f2205a0
712111f
21fd887
992268b
6fd095f
71d7764
c0f3990
73015a7
ad81682
5a521e7
daf7029
3060a2d
4be9431
8aa7919
c895440
9a9cef0
f236bc8
66cca27
7e49157
a294d56
febe9a0
83c1bd3
9469a62
90a233b
539cc86
777a580
52b7ba5
f7119ee
9186c3b
e2d4e0d
c7d4106
edc5e85
7e70a20
720f4a0
61139b9
86ab2b7
9401e87
fd467de
ee43001
1078eb3
6a7636d
5bac8c4
5d0721b
a1c0044
35c229f
4222c2e
7d91c74
01aac9c
9c3e95d
050c19d
3d484a1
229d040
53d9b60
fde77a5
af0d9dd
999f338
9581137
1591024
f2c7f34
03b454e
cefb7da
3b85915
14be623
c5851b0
65d6066
978a32c
9f52222
f5a0354
05e9788
4f6505f
e305e3a
a4f6d7a
bdd2cfd
bbef9aa
1054174
63b42b9
6f00408
055694c
90842d0
5c96627
c606fb2
e2dfc21
db90dbb
77c5e49
5092051
973f642
c1bd784
dc3afdd
4ca984a
47b850e
b7433ab
2d21015
47f85d1
cb32d0b
1c54cf7
da0b33f
b0afa4f
9d74bc6
7ee2037
3998ccb
70afe20
f4caccd
25de5bf
868fb88
18839b0
1f375f0
de69b21
4fc87fa
d76951b
73567db
6b22e36
6563824
2821164
47066b2
9e6661c
6969ff2
bb021a5
8a82f84
c29e1ae
5b4d801
e77f0b0
3308056
167c9b8
d4d0dca
af54614
bf3bfdc
0c542e9
45ea433
fec4d59
812e1ce
ea5c962
4650730
584fed8
f959966
20b3d42
a522578
98c26bc
4eebd73
9df3bbb
dcf18b1
e45d4b7
adc4708
3d49f77
afc4293
44b4642
4468692
b117e81
c164e20
c6e9dfa
3c962b0
7e2025c
273aa2c
cba3233
7413f0f
e18429c
f770113
e3463ad
5bec24e
035092f
62d72e7
eb1e452
d318627
0b780fe
a68276d
78238ba
08cee8f
6c9e183
ed5dc41
5b21de6
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
Started at: 17:01:49
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Started at: 17:03:05
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Started at: 17:03:18
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 1.5561567282676696 - MAE: 1.1770511100698875
Validation loss : 0.939958651860555 - MAE: 0.8876047142856601
Epoch: 1
Training loss: 0.6363586246967315 - MAE: 0.6928943697649724
Validation loss : 0.33423560195498997 - MAE: 0.45375851608477785
Epoch: 2
Training loss: 0.2759279823303223 - MAE: 0.4052092643476477
Validation loss : 0.20369169281588662 - MAE: 0.3433357920074413
Epoch: 3
Training loss: 0.20272068321704864 - MAE: 0.3399617345503301
Validation loss : 0.19545602964030373 - MAE: 0.3416856898288076
Epoch: 4
Training loss: 0.1912100625038147 - MAE: 0.3326939730215834
Validation loss : 0.19473741783036125 - MAE: 0.3417824430892937
Epoch: 5
Training loss: 0.19083771824836732 - MAE: 0.33137282498844967
Validation loss : 0.19215407967567444 - MAE: 0.33912431491105655
Epoch: 6
Training loss: 0.18923945665359498 - MAE: 0.33146465315973606
Validation loss : 0.18989400565624237 - MAE: 0.33674625605103725
Epoch: 7
Training loss: 0.18950583219528197 - MAE: 0.33087934479225867
Validation loss : 0.18789740900198618 - MAE: 0.33464578460452915
Epoch: 8
Training loss: 0.1856936401128769 - MAE: 0.3275132613322806
Validation loss : 0.1864797506067488 - MAE: 0.33324104491047596
Epoch: 9
Training loss: 0.1836453753709793 - MAE: 0.3255223596708951
Validation loss : 0.18501530918810102 - MAE: 0.3316937400385192
Epoch: 10
Training loss: 0.18246697008609772 - MAE: 0.32322550938186817
Validation loss : 0.18376492957274118 - MAE: 0.33042730021623856
Epoch: 11
Training loss: 0.1811656492948532 - MAE: 0.32419844121595726
Validation loss : 0.18262643449836308 - MAE: 0.32929917485750665
Epoch: 12
Training loss: 0.18094802320003509 - MAE: 0.32277421430430436
Validation loss : 0.18164582550525665 - MAE: 0.3283432280968596
Epoch: 13
Training loss: 0.1792392772436142 - MAE: 0.3209453182214294
Validation loss : 0.18065459695127276 - MAE: 0.3273255372779676
Epoch: 14
Training loss: 0.17940414905548097 - MAE: 0.3216579290630419
Validation loss : 0.17975443104902902 - MAE: 0.3264252069830345
Epoch: 15
Training loss: 0.17663683056831359 - MAE: 0.31768596289501483
Validation loss : 0.17874346673488617 - MAE: 0.3253559679886541
Epoch: 16
Training loss: 0.17649461567401886 - MAE: 0.3184662269845276
Validation loss : 0.17812747756640115 - MAE: 0.32483073473627894
Epoch: 17
Training loss: 0.17691377580165862 - MAE: 0.31822431281431196
Validation loss : 0.17722326185968187 - MAE: 0.3238687376329823
Epoch: 18
Training loss: 0.1768798279762268 - MAE: 0.3188422584293619
Validation loss : 0.17655367155869803 - MAE: 0.32321426470152304
Epoch: 19
Training loss: 0.17445733308792113 - MAE: 0.31640191359506836
Validation loss : 0.17601815693908268 - MAE: 0.3227230412018289
Epoch: 20
Training loss: 0.17480538487434388 - MAE: 0.3161891511192775
Validation loss : 0.1753774748908149 - MAE: 0.32208324574836594
Epoch: 21
Training loss: 0.1732891607284546 - MAE: 0.31558897873451675
Validation loss : 0.1746254563331604 - MAE: 0.3213051447966539
Epoch: 22
Training loss: 0.17241579234600068 - MAE: 0.3140265985892311
Validation loss : 0.17429026464621225 - MAE: 0.3210719014371899
Epoch: 23
Training loss: 0.17162418186664583 - MAE: 0.31408943363705194
Validation loss : 0.17367220256063673 - MAE: 0.32039953149205447
Epoch: 24
Training loss: 0.17133045554161072 - MAE: 0.3127541530123475
Validation loss : 0.17310193015469444 - MAE: 0.31980652640276674
Epoch: 25
Training loss: 0.16942350268363954 - MAE: 0.3121290878409379
Validation loss : 0.17268804874685076 - MAE: 0.31942688696053373
Epoch: 26
Training loss: 0.17049773633480073 - MAE: 0.31363952840493803
Validation loss : 0.172120107544793 - MAE: 0.31881123537483563
Epoch: 27
Training loss: 0.16871756613254546 - MAE: 0.3115901409937759
Validation loss : 0.17177432609928978 - MAE: 0.31851900529837285
Epoch: 28
Training loss: 0.16907432436943054 - MAE: 0.31170754064878
Validation loss : 0.17153260277377236 - MAE: 0.31832788132070616
Epoch: 29
Training loss: 0.16863942086696626 - MAE: 0.30940055904153907
Validation loss : 0.17075370252132416 - MAE: 0.31735976568837887
Epoch: 30
Training loss: 0.16562108039855958 - MAE: 0.3078162592146858
Validation loss : 0.17026153869099087 - MAE: 0.3168285940423929
Epoch: 31
Training loss: 0.16831409573554992 - MAE: 0.3108275862555087
Validation loss : 0.17004851169056362 - MAE: 0.3166782612622688
Epoch: 32
Training loss: 0.16607610881328583 - MAE: 0.3073008856775552
Validation loss : 0.169812281926473 - MAE: 0.3164999426442513
Epoch: 33
Training loss: 0.16536492824554444 - MAE: 0.30730756893246436
Validation loss : 0.16954073972172207 - MAE: 0.3162310921518069
Epoch: 34
Training loss: 0.16565500557422638 - MAE: 0.30880801031262106
Validation loss : 0.16896353165308634 - MAE: 0.3154897166767324
Epoch: 35
Training loss: 0.16480786919593812 - MAE: 0.3066865369170677
Validation loss : 0.16894808411598206 - MAE: 0.315587755169499
Epoch: 36
Training loss: 0.16488760769367217 - MAE: 0.3062544913178101
Validation loss : 0.1683160811662674 - MAE: 0.3147328218692977
Epoch: 37
Training loss: 0.16296948552131651 - MAE: 0.30529006908692957
Validation loss : 0.16822955012321472 - MAE: 0.31474619083221317
Epoch: 38
Training loss: 0.16405799746513366 - MAE: 0.3065666077277542
Validation loss : 0.16793323556582132 - MAE: 0.3143827065817725
Epoch: 39
Training loss: 0.16395046710968017 - MAE: 0.3061178662220702
Validation loss : 0.16756848825348747 - MAE: 0.31392760086070465
Epoch: 40
Training loss: 0.16225593984127046 - MAE: 0.30391575924794567
Validation loss : 0.16728156473901537 - MAE: 0.31359587016220536
Epoch: 41
Training loss: 0.162018803358078 - MAE: 0.3050126817417
Validation loss : 0.16726960241794586 - MAE: 0.3136659181800001
Epoch: 42
Training loss: 0.1616269725561142 - MAE: 0.3047072823079763
Validation loss : 0.1666801455948088 - MAE: 0.31286053951971926
Epoch: 43
Training loss: 0.16371312737464905 - MAE: 0.3072663025489704
Validation loss : 0.16660702228546143 - MAE: 0.3128544246563194
Epoch: 44
Training loss: 0.16274581968784332 - MAE: 0.3055276315731144
Validation loss : 0.16622301439444223 - MAE: 0.3123737118409309
Epoch: 45
Training loss: 0.16198877573013307 - MAE: 0.30387318215935566
Validation loss : 0.16592974132961696 - MAE: 0.3120266870766228
Epoch: 46
Training loss: 0.1613679423928261 - MAE: 0.30285112156406757
Validation loss : 0.16592562860912746 - MAE: 0.3120895779099974
Epoch: 47
Training loss: 0.1596740049123764 - MAE: 0.30196700285423644
Validation loss : 0.1656825906700558 - MAE: 0.3117942734701369
Epoch: 48
Training loss: 0.16010645925998687 - MAE: 0.3010781145976006
Validation loss : 0.1654240306880739 - MAE: 0.3114403430988788
Epoch: 49
Training loss: 0.16135582566261292 - MAE: 0.3033602255531342
Validation loss : 0.16523137191931406 - MAE: 0.3111766545135443
Epoch: 50
Training loss: 0.15887635439634323 - MAE: 0.30139360974455354
Validation loss : 0.16520478659205967 - MAE: 0.31118266518707904
Epoch: 51
Training loss: 0.1623841804265976 - MAE: 0.3041726753025802
Validation loss : 0.16488962206575605 - MAE: 0.3107705042403597
Epoch: 52
Training loss: 0.15951467990875245 - MAE: 0.3024485706726396
Validation loss : 0.16450322336620754 - MAE: 0.3101646703863994
Epoch: 53
Training loss: 0.1593264466524124 - MAE: 0.3019928035241633
Validation loss : 0.16450072824954987 - MAE: 0.3102381133991464
Epoch: 54
Training loss: 0.15892296969890596 - MAE: 0.3010730142919314
Validation loss : 0.1644514716333813 - MAE: 0.31025052996699415
Epoch: 55
Training loss: 0.15801858335733412 - MAE: 0.300129276806219
Validation loss : 0.1639546040031645 - MAE: 0.30954432283469774
Epoch: 56
Training loss: 0.15852347761392593 - MAE: 0.30000137034695895
Validation loss : 0.16394910713036856 - MAE: 0.3096252551544829
Epoch: 57
Training loss: 0.15989623069763184 - MAE: 0.3008597511380722
Validation loss : 0.1637496915128496 - MAE: 0.30938490975556004
Epoch: 58
Training loss: 0.15900149762630464 - MAE: 0.30091532233302887
Validation loss : 0.16357198854287466 - MAE: 0.3091647551750498
Epoch: 59
Training loss: 0.15815627008676528 - MAE: 0.2998753613511919
Validation loss : 0.16364828579955631 - MAE: 0.30934114039380045
Epoch: 60
Training loss: 0.15863654345273973 - MAE: 0.2997401767468425
Validation loss : 0.16360284553633797 - MAE: 0.3093700140804052
Epoch: 61
Training loss: 0.1569861376285553 - MAE: 0.3001724729244438
Validation loss : 0.1631663309203254 - MAE: 0.3087389482069802
Epoch: 62
Training loss: 0.15764495313167573 - MAE: 0.29931992861180673
Validation loss : 0.1628624697526296 - MAE: 0.3082940994455321
Epoch: 63
Training loss: 0.15620498329401017 - MAE: 0.2983585146384368
Validation loss : 0.16287368204858568 - MAE: 0.30839281732122203
Epoch: 64
Training loss: 0.15624627590179443 - MAE: 0.29862051694345537
Validation loss : 0.16266994840568966 - MAE: 0.308149927566448
Epoch: 65
Training loss: 0.15566111862659454 - MAE: 0.29753564913606617
Validation loss : 0.1626807086997562 - MAE: 0.30822756782090394
Epoch: 66
Training loss: 0.15647490918636323 - MAE: 0.29872063729818527
Validation loss : 0.16234126521481407 - MAE: 0.3077123779358625
Epoch: 67
Training loss: 0.15619896858930588 - MAE: 0.297556687519032
Validation loss : 0.16227295829190147 - MAE: 0.30769900980501297
Epoch: 68
Training loss: 0.15651787519454957 - MAE: 0.2983152167118014
Validation loss : 0.16212009721332127 - MAE: 0.30753654402945435
Epoch: 69
Training loss: 0.15651776432991027 - MAE: 0.299202365307033
Validation loss : 0.16212468677096897 - MAE: 0.3076282798695557
Epoch: 70
Training loss: 0.15673158675432206 - MAE: 0.29856837209150694
Validation loss : 0.16198681791623434 - MAE: 0.30741784086039886
Epoch: 71
Training loss: 0.15566783308982848 - MAE: 0.2983743266636147
Validation loss : 0.1617990599738227 - MAE: 0.3071957299555147
Epoch: 72
Training loss: 0.1555081295967102 - MAE: 0.296568577820393
Validation loss : 0.16165704197353786 - MAE: 0.30704654154909133
Epoch: 73
Training loss: 0.1559882402420044 - MAE: 0.2972687981178256