File size: 6,863 Bytes
4001542
f70081d
bfe4889
57c6f06
420046c
5e0a3da
e8ebc03
c4038c0
7e2b43b
6bc8b46
7184377
c7792bd
e70cc41
7d1a920
a947808
e8e06a9
18308cf
d083566
4b2416e
e7ba8a8
2a1d7ec
a4b8a87
08546e8
c040908
ccb9371
dd2283a
2a94064
8d19120
d0fd6fa
d971075
2f11030
00ffef7
7589303
9f7427d
215ca20
472f0c7
8e325e8
3154351
86b7e6d
b7e6a40
15b9ad9
ac7d759
8d91f1c
e249531
6875df7
34108ab
0325135
2587462
08b7760
e091124
e09f6a5
f0958b0
df09ca9
aad6266
404b115
0b31c62
907ef70
f1817ea
dbcf0fd
19b5dda
8671bd3
c85fb05
4489afc
2ad0729
874f548
452dfb0
337893d
2476c4c
e9aa6f8
27a195a
41958c8
862d359
5aef33c
0d10d1f
a96eacb
f9a157f
97d8f0d
e522da0
20c32b5
ac6fb45
c566ccf
e844f74
f61e0a0
1657b87
a5fbc8d
ef27d64
72db1b8
87233f9
dd350e2
2ef68ad
ee18590
18e3705
b124356
b24e3bf
3af3e90
da19f8b
1506dfb
678cf05
55327a6
ef20459
9a091bb
2e13c85
054204f
2fe798a
a06a528
3850eb3
3fa42ab
ff27786
1f2e6f6
77e43c9
bbd8ca8
416fd8c
6c0d046
9aac2df
dab6b41
313f18f
f1124ca
680d38d
f3ac684
b3ffc7f
828762d
0485ce5
0537688
0137ad4
884f1c6
94c9f76
18b2dee
4516998
70eb2e0
67f31f8
4e82dde
5595ae3
81b0a83
152a3e7
21fb2a3
1ca7565
213c584
83fda25
25a1929
7c75cc0
b10d97b
6999f11
829f7f7
6efe931
99b128e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
Started at: 14:58:05
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '44815f7e109b53547cccdf3c6847f4c28b989816'}, {})
Epoch: 0
Training loss: 0.25046255227726727 - MSE: 0.36727700175627953
Validation loss : 0.17648151460470576 - MSE: 0.32341695900998885
Epoch: 1
Training loss: 0.17091115193469875 - MSE: 0.31430697735027086
Validation loss : 0.16562944582917474 - MSE: 0.31155962633828027
Epoch: 2
Training loss: 0.16704481191517132 - MSE: 0.3114155539092336
Validation loss : 0.1604571777085463 - MSE: 0.30340937581848676
Epoch: 3
Training loss: 0.16466754909396777 - MSE: 0.30768393210761735
Validation loss : 0.1632813240091006 - MSE: 0.3109833563942495
Epoch: 4
Training loss: 0.16455070379421796 - MSE: 0.3093531171941453
Validation loss : 0.16564318244204376 - MSE: 0.30289557314252813
Epoch: 5
Training loss: 0.16568223224376058 - MSE: 0.30846559528561224
Validation loss : 0.1605968801022479 - MSE: 0.2988926674776124
Epoch: 6
Training loss: 0.15717696056223762 - MSE: 0.3022554811747332
Validation loss : 0.16166257779255058 - MSE: 0.29849396869512723
Epoch: 7
Training loss: 0.15822318865714338 - MSE: 0.30195677281494104
Validation loss : 0.1757645166733048 - MSE: 0.3080704490543514
Epoch: 8
Training loss: 0.15552518461939646 - MSE: 0.2998732285398632
Validation loss : 0.1911594748045459 - MSE: 0.32289788546992626
Epoch: 9
Training loss: 0.15594329739812063 - MSE: 0.3009645091966726
Validation loss : 0.19870018710692725 - MSE: 0.33138930551919304
Epoch: 10
Training loss: 0.1547534751521452 - MSE: 0.2992649989557918
Validation loss : 0.19730406333551262 - MSE: 0.3297122567265123
Epoch: 11
Training loss: 0.15529195780908395 - MSE: 0.29975488127620853
Validation loss : 0.18655535093311107 - MSE: 0.3217206286243775
Epoch: 12
Training loss: 0.15309068918001228 - MSE: 0.2969653755509521
Validation loss : 0.17314820917266788 - MSE: 0.30849219803901623
Epoch: 13
Training loss: 0.15322715030784534 - MSE: 0.29618723127423563
Validation loss : 0.1593460923794544 - MSE: 0.29643875794480956
Epoch: 14
Training loss: 0.1519091561209732 - MSE: 0.2967038231193114
Validation loss : 0.15947996328274408 - MSE: 0.2962177626613215
Epoch: 15
Training loss: 0.1513304633234963 - MSE: 0.2948141613710868
Validation loss : 0.17381140318783847 - MSE: 0.30743163511355315
Epoch: 16
Training loss: 0.15304585124559814 - MSE: 0.2953887584017137
Validation loss : 0.15291882255537945 - MSE: 0.2937530252336034
Epoch: 17
Training loss: 0.15207511008844762 - MSE: 0.29601101879974745
Validation loss : 0.15711521870936407 - MSE: 0.2956258048557993
Epoch: 18
Training loss: 0.1522078831032448 - MSE: 0.2947807228620845
Validation loss : 0.1593479738768303 - MSE: 0.2975361759408073
Epoch: 19
Training loss: 0.15030200151622597 - MSE: 0.29364889133341376
Validation loss : 0.15571271227390476 - MSE: 0.2947020680712253
Epoch: 20
Training loss: 0.14997340040990545 - MSE: 0.29321906843107093
Validation loss : 0.16053046500592522 - MSE: 0.2972562162079979
Epoch: 21
Training loss: 0.1503814266153096 - MSE: 0.29498264967501453
Validation loss : 0.18625259749365575 - MSE: 0.3214374853139268
Epoch: 22
Training loss: 0.15071523613130985 - MSE: 0.2931879129534037
Validation loss : 0.15868466793361938 - MSE: 0.29779115185817956
Epoch: 23
Training loss: 0.14925291788230088 - MSE: 0.29253552849679587
Validation loss : 0.17538098990917206 - MSE: 0.30947426368924724
Epoch: 24
Training loss: 0.15000015682526652 - MSE: 0.2927369972856095
Validation loss : 0.1700212608909968 - MSE: 0.3071491417052287
Epoch: 25
Training loss: 0.15024784067409291 - MSE: 0.29311721077485986
Validation loss : 0.19164935623606047 - MSE: 0.32203448128239803
Epoch: 26
Training loss: 0.15137434639225755 - MSE: 0.29379136046127924
Validation loss : 0.1671447205272588 - MSE: 0.30391976326317943
Epoch: 27
Training loss: 0.15017312853998943 - MSE: 0.29455603880812276
Validation loss : 0.18036871166391807 - MSE: 0.3125465653289461
Epoch: 28
Training loss: 0.14778636641762583 - MSE: 0.2905298917289279
Validation loss : 0.16160754949757547 - MSE: 0.2979193112805207
Epoch: 29
Training loss: 0.1477772275241196 - MSE: 0.2912173747535554
Validation loss : 0.17026665750326533 - MSE: 0.30567947675889745
Epoch: 30
Training loss: 0.14744956316757324 - MSE: 0.291312328070814
Validation loss : 0.16202468379880441 - MSE: 0.29700271247754034
Epoch: 31
Training loss: 0.1484179409778663 - MSE: 0.2921315750433443
Validation loss : 0.16338278116150337 - MSE: 0.2987860072168981
Epoch: 32
Training loss: 0.14738556740777142 - MSE: 0.29104356719237584
Validation loss : 0.15635579254365328 - MSE: 0.29443009978003276
Epoch: 33
Training loss: 0.148553161413839 - MSE: 0.29089201478673565
Validation loss : 0.15951512471744508 - MSE: 0.2966586790896474
Epoch: 34
Training loss: 0.15104458401650947 - MSE: 0.2934267254216827
Validation loss : 0.17196247625080022 - MSE: 0.3055353329027785
Epoch: 35
Training loss: 0.1503695050623211 - MSE: 0.29389694550707784
Validation loss : 0.16170481603705522 - MSE: 0.29835679015904104
Epoch: 36
Training loss: 0.14925730479928442 - MSE: 0.2924833646517312
Validation loss : 0.17856141236243825 - MSE: 0.31225271412806316
Epoch: 37
Training loss: 0.15118206122199895 - MSE: 0.29395904130187084
Validation loss : 0.17895790042750764 - MSE: 0.3121755853923419
Epoch: 38
Training loss: 0.1497652710505246 - MSE: 0.2912881449244226
Validation loss : 0.15206407851567774 - MSE: 0.2909835362196566
Epoch: 39
Training loss: 0.14738748415472544 - MSE: 0.29036707520169275
Validation loss : 0.16739020203099106 - MSE: 0.30124313077212783
Epoch: 40
Training loss: 0.14675953452886664 - MSE: 0.2900500768121404
Validation loss : 0.17623373211333246 - MSE: 0.30992423744210557
Epoch: 41
Training loss: 0.14824477195512825 - MSE: 0.29090776402279417
Validation loss : 0.16109058317361455 - MSE: 0.2974189646320188
Epoch: 42
Training loss: 0.14692961365967838 - MSE: 0.2904768427610488
Validation loss : 0.1707474204401175 - MSE: 0.30493211675533977
Epoch: 43
Training loss: 0.14761571415941122 - MSE: 0.2907837329648465
Validation loss : 0.1536927172970591 - MSE: 0.2910231603894442
Epoch: 44
Training loss: 0.15164254637917285 - MSE: 0.2942169462116838
Validation loss : 0.15892887950846643 - MSE: 0.2960443273887821
Epoch: 45
Training loss: 0.14786733518624065 - MSE: 0.2917048977763665
Validation loss : 0.1669148841139042 - MSE: 0.30125252101874894
Epoch: 46
Training loss: 0.14599600882472727 - MSE: 0.286971878648382
Validation loss : 0.15723810294135052 - MSE: 0.29333694749044353
Epoch: 47
Training loss: 0.14816145228250377 - MSE: 0.28941392496829527