File size: 2,665 Bytes
0f4ddf6
94c1893
66c6131
3b94107
fbf603f
56fe2d3
1e73aec
20f5ed8
bc9be57
86c79cd
607bfff
1783dd2
4ee6ab4
d075457
9c99b98
4353f90
8845df1
2833e42
9acac77
fdc5a9f
63ed4ae
e9d9b2d
0f5241d
4a87de6
5fe2307
2e43549
8b0f13e
6a992f2
670fd0d
4c2f141
dc71a12
e123fbc
325fe9d
fd3c27c
4c9695c
20ca70e
8d4c8f6
c19c625
9b84df8
52617dc
cf34feb
76518ab
69c74bc
78f22ab
54967e1
8b8f5fe
20d4375
7979e9f
4a0e93d
38437fd
33f3daf
950da3c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
Started at: 15:20:57
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.3436830922961235 - MAE: 0.4572523557050669
Validation loss : 0.26256299018859863 - MAE: 0.40728324728553433
Epoch: 1
Training loss: 0.21011242344975473 - MAE: 0.35190288965025124
Validation loss : 0.18477652966976166 - MAE: 0.3296733541388515
Epoch: 2
Training loss: 0.19629175141453742 - MAE: 0.33688825701191527
Validation loss : 0.18152975397450583 - MAE: 0.32990742521562766
Epoch: 3
Training loss: 0.1831872671842575 - MAE: 0.3247871467745211
Validation loss : 0.17427935344832285 - MAE: 0.3213832702684575
Epoch: 4
Training loss: 0.1773839347064495 - MAE: 0.31992620730258825
Validation loss : 0.17429945511477335 - MAE: 0.3230458382996596
Epoch: 5
Training loss: 0.17231015339493752 - MAE: 0.31171345188658794
Validation loss : 0.17113730950014933 - MAE: 0.31939983255844806
Epoch: 6
Training loss: 0.16903775408864022 - MAE: 0.3102792424124377
Validation loss : 0.16909515857696533 - MAE: 0.3172391366407421
Epoch: 7
Training loss: 0.16602707877755166 - MAE: 0.3075343290458685
Validation loss : 0.16411419212818146 - MAE: 0.31039227117874574
Epoch: 8
Training loss: 0.1662418097257614 - MAE: 0.3082071055792662
Validation loss : 0.16296945938042232 - MAE: 0.3092928071837739
Epoch: 9
Training loss: 0.16653021723031997 - MAE: 0.30672722178327916
Validation loss : 0.16100754695279257 - MAE: 0.30576533515417514
Epoch: 10
Training loss: 0.16153860688209534 - MAE: 0.30234728984229753
Validation loss : 0.16281781026295253 - MAE: 0.3046462940190389
Epoch: 11
Training loss: 0.16209525130689145 - MAE: 0.3039236849786875
Validation loss : 0.1614038199186325 - MAE: 0.30358997656721776
Epoch: 12
Training loss: 0.15954031608998775 - MAE: 0.3032080887111098
Validation loss : 0.16037237431321824 - MAE: 0.3024962801353889
Epoch: 13
Training loss: 0.1593211904168129 - MAE: 0.30014842639769856
Validation loss : 0.1610377154179982 - MAE: 0.30269789015625514
Epoch: 14
Training loss: 0.15742682591080664 - MAE: 0.2990433643869133
Validation loss : 0.16269906503813608 - MAE: 0.3034925004658223
Epoch: 15
Training loss: 0.15586433932185173 - MAE: 0.29719735232655653
Validation loss : 0.1581807051386152 - MAE: 0.3014812709966527
Epoch: 16
Training loss: 0.15324377864599228 - MAE: 0.29518426166093065