File size: 4,107 Bytes
0f4ddf6
94c1893
66c6131
3b94107
fbf603f
56fe2d3
1e73aec
20f5ed8
bc9be57
86c79cd
607bfff
1783dd2
4ee6ab4
d075457
9c99b98
4353f90
8845df1
2833e42
9acac77
fdc5a9f
63ed4ae
e9d9b2d
0f5241d
4a87de6
5fe2307
2e43549
8b0f13e
6a992f2
670fd0d
4c2f141
dc71a12
e123fbc
325fe9d
fd3c27c
4c9695c
20ca70e
8d4c8f6
c19c625
9b84df8
52617dc
cf34feb
76518ab
69c74bc
78f22ab
54967e1
8b8f5fe
20d4375
7979e9f
4a0e93d
38437fd
33f3daf
950da3c
705a3dc
03b6e28
7b0bbd4
334f1aa
d4299e0
a15807b
fb9db54
bd463e7
4e15208
48e1193
e868138
f3e47f4
394f243
c369ae7
440aa5c
ed96971
2d9c75e
fe11629
abd7d78
2595a0e
21a9c8f
caf392f
62192e6
671a46a
dfc3aed
b1437fe
73719d1
7b64bb1
6b42eb7
3a93cd6
a90ce90
7a098dc
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
Started at: 15:20:57
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.3436830922961235 - MAE: 0.4572523557050669
Validation loss : 0.26256299018859863 - MAE: 0.40728324728553433
Epoch: 1
Training loss: 0.21011242344975473 - MAE: 0.35190288965025124
Validation loss : 0.18477652966976166 - MAE: 0.3296733541388515
Epoch: 2
Training loss: 0.19629175141453742 - MAE: 0.33688825701191527
Validation loss : 0.18152975397450583 - MAE: 0.32990742521562766
Epoch: 3
Training loss: 0.1831872671842575 - MAE: 0.3247871467745211
Validation loss : 0.17427935344832285 - MAE: 0.3213832702684575
Epoch: 4
Training loss: 0.1773839347064495 - MAE: 0.31992620730258825
Validation loss : 0.17429945511477335 - MAE: 0.3230458382996596
Epoch: 5
Training loss: 0.17231015339493752 - MAE: 0.31171345188658794
Validation loss : 0.17113730950014933 - MAE: 0.31939983255844806
Epoch: 6
Training loss: 0.16903775408864022 - MAE: 0.3102792424124377
Validation loss : 0.16909515857696533 - MAE: 0.3172391366407421
Epoch: 7
Training loss: 0.16602707877755166 - MAE: 0.3075343290458685
Validation loss : 0.16411419212818146 - MAE: 0.31039227117874574
Epoch: 8
Training loss: 0.1662418097257614 - MAE: 0.3082071055792662
Validation loss : 0.16296945938042232 - MAE: 0.3092928071837739
Epoch: 9
Training loss: 0.16653021723031997 - MAE: 0.30672722178327916
Validation loss : 0.16100754695279257 - MAE: 0.30576533515417514
Epoch: 10
Training loss: 0.16153860688209534 - MAE: 0.30234728984229753
Validation loss : 0.16281781026295253 - MAE: 0.3046462940190389
Epoch: 11
Training loss: 0.16209525130689145 - MAE: 0.3039236849786875
Validation loss : 0.1614038199186325 - MAE: 0.30358997656721776
Epoch: 12
Training loss: 0.15954031608998775 - MAE: 0.3032080887111098
Validation loss : 0.16037237431321824 - MAE: 0.3024962801353889
Epoch: 13
Training loss: 0.1593211904168129 - MAE: 0.30014842639769856
Validation loss : 0.1610377154179982 - MAE: 0.30269789015625514
Epoch: 14
Training loss: 0.15742682591080664 - MAE: 0.2990433643869133
Validation loss : 0.16269906503813608 - MAE: 0.3034925004658223
Epoch: 15
Training loss: 0.15586433932185173 - MAE: 0.29719735232655653
Validation loss : 0.1581807051386152 - MAE: 0.3014812709966527
Epoch: 16
Training loss: 0.15324377864599228 - MAE: 0.29518426166093065
Validation loss : 0.15657826832362584 - MAE: 0.300699633359451
Epoch: 17
Training loss: 0.15139634534716606 - MAE: 0.2926187599905815
Validation loss : 0.15587361582687922 - MAE: 0.2989909241765162
Epoch: 18
Training loss: 0.150983314961195 - MAE: 0.2925825479179498
Validation loss : 0.15516561269760132 - MAE: 0.29856372533592385
Epoch: 19
Training loss: 0.1493927862495184 - MAE: 0.2910948114445842
Validation loss : 0.15466026110308512 - MAE: 0.2976538621938452
Epoch: 20
Training loss: 0.14843332655727864 - MAE: 0.2885189939815885
Validation loss : 0.15413911640644073 - MAE: 0.29834344421420217
Epoch: 21
Training loss: 0.1489902339875698 - MAE: 0.2893552038446195
Validation loss : 0.15573792159557343 - MAE: 0.29749915735213595
Epoch: 22
Training loss: 0.14985780157148837 - MAE: 0.29154637065109446
Validation loss : 0.15332231564181192 - MAE: 0.29697695339866076
Epoch: 23
Training loss: 0.14575221315026282 - MAE: 0.2886255141256018
Validation loss : 0.15240811024393355 - MAE: 0.29550651769807773
Epoch: 24
Training loss: 0.14468684382736682 - MAE: 0.2861701193601813
Validation loss : 0.15227091951029642 - MAE: 0.2969623801494556
Epoch: 25
Training loss: 0.14725722186267376 - MAE: 0.28829223441788526
Validation loss : 0.15270294461931502 - MAE: 0.29577565945239215
Epoch: 26
Training loss: 0.14587189629673958 - MAE: 0.2862314013226889
Validation loss : 0.1532691674573081 - MAE: 0.29749263015344235
Stopped after 26 epochs.