File size: 8,201 Bytes
ee53a4a
50f1745
8a33155
79fb205
22be4a5
3f1ff0e
bbed904
14c983c
bb23377
0ea8477
8597b5f
a529489
8f5752b
d444cf9
2dfa794
782f4ba
2665515
8c5051b
84a2460
36464be
b4eab8a
b4c7480
bd5a96e
c248496
a3f3665
3572268
8f95e8a
0e3f82d
bd48add
1da9fd2
3759f70
e09bf1b
ec4c3b8
b891788
9c846de
c36a8be
3f890ba
aaf9dcb
9442814
64e656f
47314ea
b48fff4
c931c86
ec8e58d
28782de
d61067e
2cde297
dce64ed
1a44567
a894d88
f8b0b44
63524f2
3dc095f
915dd82
a5a47b9
15b72bc
5ba086a
6a406c0
fd9bd7c
f2205a0
712111f
21fd887
992268b
6fd095f
71d7764
c0f3990
73015a7
ad81682
5a521e7
daf7029
3060a2d
4be9431
8aa7919
c895440
9a9cef0
f236bc8
66cca27
7e49157
a294d56
febe9a0
83c1bd3
9469a62
90a233b
539cc86
777a580
52b7ba5
f7119ee
9186c3b
e2d4e0d
c7d4106
edc5e85
7e70a20
720f4a0
61139b9
86ab2b7
9401e87
fd467de
ee43001
1078eb3
6a7636d
5bac8c4
5d0721b
a1c0044
35c229f
4222c2e
7d91c74
01aac9c
9c3e95d
050c19d
3d484a1
229d040
53d9b60
fde77a5
af0d9dd
999f338
9581137
1591024
f2c7f34
03b454e
cefb7da
3b85915
14be623
c5851b0
65d6066
978a32c
9f52222
f5a0354
05e9788
4f6505f
e305e3a
a4f6d7a
bdd2cfd
bbef9aa
1054174
63b42b9
6f00408
055694c
90842d0
5c96627
c606fb2
e2dfc21
db90dbb
77c5e49
5092051
973f642
c1bd784
dc3afdd
4ca984a
47b850e
b7433ab
2d21015
47f85d1
cb32d0b
1c54cf7
da0b33f
b0afa4f
9d74bc6
7ee2037
3998ccb
70afe20
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
Started at: 17:01:49
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Started at: 17:03:05
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Started at: 17:03:18
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 1.5561567282676696 - MAE: 1.1770511100698875
Validation loss : 0.939958651860555 - MAE: 0.8876047142856601
Epoch: 1
Training loss: 0.6363586246967315 - MAE: 0.6928943697649724
Validation loss : 0.33423560195498997 - MAE: 0.45375851608477785
Epoch: 2
Training loss: 0.2759279823303223 - MAE: 0.4052092643476477
Validation loss : 0.20369169281588662 - MAE: 0.3433357920074413
Epoch: 3
Training loss: 0.20272068321704864 - MAE: 0.3399617345503301
Validation loss : 0.19545602964030373 - MAE: 0.3416856898288076
Epoch: 4
Training loss: 0.1912100625038147 - MAE: 0.3326939730215834
Validation loss : 0.19473741783036125 - MAE: 0.3417824430892937
Epoch: 5
Training loss: 0.19083771824836732 - MAE: 0.33137282498844967
Validation loss : 0.19215407967567444 - MAE: 0.33912431491105655
Epoch: 6
Training loss: 0.18923945665359498 - MAE: 0.33146465315973606
Validation loss : 0.18989400565624237 - MAE: 0.33674625605103725
Epoch: 7
Training loss: 0.18950583219528197 - MAE: 0.33087934479225867
Validation loss : 0.18789740900198618 - MAE: 0.33464578460452915
Epoch: 8
Training loss: 0.1856936401128769 - MAE: 0.3275132613322806
Validation loss : 0.1864797506067488 - MAE: 0.33324104491047596
Epoch: 9
Training loss: 0.1836453753709793 - MAE: 0.3255223596708951
Validation loss : 0.18501530918810102 - MAE: 0.3316937400385192
Epoch: 10
Training loss: 0.18246697008609772 - MAE: 0.32322550938186817
Validation loss : 0.18376492957274118 - MAE: 0.33042730021623856
Epoch: 11
Training loss: 0.1811656492948532 - MAE: 0.32419844121595726
Validation loss : 0.18262643449836308 - MAE: 0.32929917485750665
Epoch: 12
Training loss: 0.18094802320003509 - MAE: 0.32277421430430436
Validation loss : 0.18164582550525665 - MAE: 0.3283432280968596
Epoch: 13
Training loss: 0.1792392772436142 - MAE: 0.3209453182214294
Validation loss : 0.18065459695127276 - MAE: 0.3273255372779676
Epoch: 14
Training loss: 0.17940414905548097 - MAE: 0.3216579290630419
Validation loss : 0.17975443104902902 - MAE: 0.3264252069830345
Epoch: 15
Training loss: 0.17663683056831359 - MAE: 0.31768596289501483
Validation loss : 0.17874346673488617 - MAE: 0.3253559679886541
Epoch: 16
Training loss: 0.17649461567401886 - MAE: 0.3184662269845276
Validation loss : 0.17812747756640115 - MAE: 0.32483073473627894
Epoch: 17
Training loss: 0.17691377580165862 - MAE: 0.31822431281431196
Validation loss : 0.17722326185968187 - MAE: 0.3238687376329823
Epoch: 18
Training loss: 0.1768798279762268 - MAE: 0.3188422584293619
Validation loss : 0.17655367155869803 - MAE: 0.32321426470152304
Epoch: 19
Training loss: 0.17445733308792113 - MAE: 0.31640191359506836
Validation loss : 0.17601815693908268 - MAE: 0.3227230412018289
Epoch: 20
Training loss: 0.17480538487434388 - MAE: 0.3161891511192775
Validation loss : 0.1753774748908149 - MAE: 0.32208324574836594
Epoch: 21
Training loss: 0.1732891607284546 - MAE: 0.31558897873451675
Validation loss : 0.1746254563331604 - MAE: 0.3213051447966539
Epoch: 22
Training loss: 0.17241579234600068 - MAE: 0.3140265985892311
Validation loss : 0.17429026464621225 - MAE: 0.3210719014371899
Epoch: 23
Training loss: 0.17162418186664583 - MAE: 0.31408943363705194
Validation loss : 0.17367220256063673 - MAE: 0.32039953149205447
Epoch: 24
Training loss: 0.17133045554161072 - MAE: 0.3127541530123475
Validation loss : 0.17310193015469444 - MAE: 0.31980652640276674
Epoch: 25
Training loss: 0.16942350268363954 - MAE: 0.3121290878409379
Validation loss : 0.17268804874685076 - MAE: 0.31942688696053373
Epoch: 26
Training loss: 0.17049773633480073 - MAE: 0.31363952840493803
Validation loss : 0.172120107544793 - MAE: 0.31881123537483563
Epoch: 27
Training loss: 0.16871756613254546 - MAE: 0.3115901409937759
Validation loss : 0.17177432609928978 - MAE: 0.31851900529837285
Epoch: 28
Training loss: 0.16907432436943054 - MAE: 0.31170754064878
Validation loss : 0.17153260277377236 - MAE: 0.31832788132070616
Epoch: 29
Training loss: 0.16863942086696626 - MAE: 0.30940055904153907
Validation loss : 0.17075370252132416 - MAE: 0.31735976568837887
Epoch: 30
Training loss: 0.16562108039855958 - MAE: 0.3078162592146858
Validation loss : 0.17026153869099087 - MAE: 0.3168285940423929
Epoch: 31
Training loss: 0.16831409573554992 - MAE: 0.3108275862555087
Validation loss : 0.17004851169056362 - MAE: 0.3166782612622688
Epoch: 32
Training loss: 0.16607610881328583 - MAE: 0.3073008856775552
Validation loss : 0.169812281926473 - MAE: 0.3164999426442513
Epoch: 33
Training loss: 0.16536492824554444 - MAE: 0.30730756893246436
Validation loss : 0.16954073972172207 - MAE: 0.3162310921518069
Epoch: 34
Training loss: 0.16565500557422638 - MAE: 0.30880801031262106
Validation loss : 0.16896353165308634 - MAE: 0.3154897166767324
Epoch: 35
Training loss: 0.16480786919593812 - MAE: 0.3066865369170677
Validation loss : 0.16894808411598206 - MAE: 0.315587755169499
Epoch: 36
Training loss: 0.16488760769367217 - MAE: 0.3062544913178101
Validation loss : 0.1683160811662674 - MAE: 0.3147328218692977
Epoch: 37
Training loss: 0.16296948552131651 - MAE: 0.30529006908692957
Validation loss : 0.16822955012321472 - MAE: 0.31474619083221317
Epoch: 38
Training loss: 0.16405799746513366 - MAE: 0.3065666077277542
Validation loss : 0.16793323556582132 - MAE: 0.3143827065817725
Epoch: 39
Training loss: 0.16395046710968017 - MAE: 0.3061178662220702
Validation loss : 0.16756848825348747 - MAE: 0.31392760086070465
Epoch: 40
Training loss: 0.16225593984127046 - MAE: 0.30391575924794567
Validation loss : 0.16728156473901537 - MAE: 0.31359587016220536
Epoch: 41
Training loss: 0.162018803358078 - MAE: 0.3050126817417
Validation loss : 0.16726960241794586 - MAE: 0.3136659181800001
Epoch: 42
Training loss: 0.1616269725561142 - MAE: 0.3047072823079763
Validation loss : 0.1666801455948088 - MAE: 0.31286053951971926
Epoch: 43
Training loss: 0.16371312737464905 - MAE: 0.3072663025489704
Validation loss : 0.16660702228546143 - MAE: 0.3128544246563194
Epoch: 44
Training loss: 0.16274581968784332 - MAE: 0.3055276315731144
Validation loss : 0.16622301439444223 - MAE: 0.3123737118409309
Epoch: 45
Training loss: 0.16198877573013307 - MAE: 0.30387318215935566
Validation loss : 0.16592974132961696 - MAE: 0.3120266870766228
Epoch: 46
Training loss: 0.1613679423928261 - MAE: 0.30285112156406757
Validation loss : 0.16592562860912746 - MAE: 0.3120895779099974
Epoch: 47
Training loss: 0.1596740049123764 - MAE: 0.30196700285423644
Validation loss : 0.1656825906700558 - MAE: 0.3117942734701369
Epoch: 48
Training loss: 0.16010645925998687 - MAE: 0.3010781145976006
Validation loss : 0.1654240306880739 - MAE: 0.3114403430988788
Epoch: 49
Training loss: 0.16135582566261292 - MAE: 0.3033602255531342
Validation loss : 0.16523137191931406 - MAE: 0.3111766545135443
Epoch: 50
Training loss: 0.15887635439634323 - MAE: 0.30139360974455354
Validation loss : 0.16520478659205967 - MAE: 0.31118266518707904