File size: 2,911 Bytes
fd1f392
3818fca
74d56c4
38dbb06
170c32e
e54303a
6d84434
fb53ab9
f2eea60
a955de2
3c17317
e344626
a2fa97c
180c5c1
f8ae289
4ed60da
581031b
ff1c738
770c353
dcce380
e4d03fe
30fe58f
0b11992
13127f0
2606fa1
3fc53eb
125e149
cb46dd4
19bb37b
a190d71
c54bf5c
a33fa5a
d1ef7d5
1aab339
e575809
1fa4dbb
cf93c8f
0424846
87fe95e
e144505
5c44a5e
79d3460
f8f3f82
16c5678
407a522
b64fc47
c38e4cc
bcc6b65
88ae8ac
f10daa8
76a0720
9437d18
e19f589
a7eab35
433b3b3
a6fe8ca
8b3d5ea
b223896
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
Started at: 11:12:13
norbert2, 5e-06, 256
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.41463836485689337 - MAE: 0.5087427320675125
Validation loss : 0.2686559503015719 - MAE: 0.4043045732741782
Epoch: 1
Training loss: 0.24431555514985864 - MAE: 0.38740748680688086
Validation loss : 0.18667236833195938 - MAE: 0.33838491992877257
Epoch: 2
Training loss: 0.18506982191042468 - MAE: 0.3382457084132687
Validation loss : 0.15371774607583097 - MAE: 0.3074420312338156
Epoch: 3
Training loss: 0.1546944863416932 - MAE: 0.31008215112442716
Validation loss : 0.133816171241434 - MAE: 0.2858432389366333
Epoch: 4
Training loss: 0.13350679468024862 - MAE: 0.28923511177879146
Validation loss : 0.10932125621720364 - MAE: 0.256454776400781
Epoch: 5
Training loss: 0.1168869366699999 - MAE: 0.26986456802804265
Validation loss : 0.09649829840973805 - MAE: 0.23724551405129826
Epoch: 6
Training loss: 0.10431941856037487 - MAE: 0.2534326974131011
Validation loss : 0.08749974440587194 - MAE: 0.22284109841832134
Epoch: 7
Training loss: 0.08919526636600494 - MAE: 0.2336376825413137
Validation loss : 0.07944989831824052 - MAE: 0.2086082506448854
Epoch: 8
Training loss: 0.08102413564920426 - MAE: 0.22234788799560584
Validation loss : 0.07408433172263597 - MAE: 0.19957130841128806
Epoch: 9
Training loss: 0.07282835455103354 - MAE: 0.20905587992288474
Validation loss : 0.07321358158400185 - MAE: 0.1958791246236593
Epoch: 10
Training loss: 0.06805636808276176 - MAE: 0.2019703178512447
Validation loss : 0.06818335130810738 - MAE: 0.18784236918567274
Epoch: 11
Training loss: 0.06315255009315231 - MAE: 0.19438868563685022
Validation loss : 0.061382615840748736 - MAE: 0.17627332670686088
Epoch: 12
Training loss: 0.05669153705239296 - MAE: 0.18379919555202343
Validation loss : 0.06136786408330265 - MAE: 0.17450401481941827
Epoch: 13
Training loss: 0.05257220755923878 - MAE: 0.17699028630098032
Validation loss : 0.060792625734680576 - MAE: 0.17298781770388472
Epoch: 14
Training loss: 0.04955695366317576 - MAE: 0.17188387365707222
Validation loss : 0.059647708543037116 - MAE: 0.17080333908403275
Epoch: 15
Training loss: 0.047338411550630226 - MAE: 0.16792417539661747
Validation loss : 0.05730088721764715 - MAE: 0.16665574399906258
Epoch: 16
Training loss: 0.044937838884917175 - MAE: 0.1632461903660146
Validation loss : 0.05832104149617647 - MAE: 0.16912266812807178
Epoch: 17
Training loss: 0.04221029823476618 - MAE: 0.15887206312399307
Validation loss : 0.056816791037195606 - MAE: 0.16741534462301444
Epoch: 18