File size: 2,859 Bytes
a54743f
da02fe8
64c7ac4
ef2b635
0bab893
743bf2f
6e0c0fe
4e0ce49
07a14de
b58387d
cea0a2c
8be760f
6987d12
e76068d
f8f2c9f
73ef6e7
db96ccb
51b5543
5f313bc
fb33851
c47912b
33fd113
5d2cc42
6a2936c
e4284ca
b7b2d5b
0128a2d
787d203
25cb336
f03d817
d6c4c1c
2626bc6
500c492
f8d559c
23aa4c6
a45ac6b
70335cf
dd9125d
c2df8a7
636aeee
8900ea4
dd55b82
aaa6ec3
4a4d811
0edfeff
4d79fd7
1392ae5
18ae2a0
5c3dc1f
726f346
0b4ddac
7586150
8804e7a
6a07917
25f5dfa
646bf85
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
Started at: 16:30:49
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 0.3458641168475151 - MAE: 0.4470721267008891
Validation loss : 0.1788818021030987 - MAE: 0.32197414947064035
Epoch: 1
Training loss: 0.17300812512636185 - MAE: 0.3175820470114712
Validation loss : 0.16896727418198304 - MAE: 0.31332903775222964
Epoch: 2
Training loss: 0.1648126335442066 - MAE: 0.308156299760969
Validation loss : 0.16454754506840424 - MAE: 0.3087028804974932
Epoch: 3
Training loss: 0.1608002531528473 - MAE: 0.3038113721048132
Validation loss : 0.1621274115408168 - MAE: 0.30596198775094147
Epoch: 4
Training loss: 0.15862647637724878 - MAE: 0.3014963360894156
Validation loss : 0.16059190341654947 - MAE: 0.30361069127564794
Epoch: 5
Training loss: 0.15564483717083932 - MAE: 0.29912276695755136
Validation loss : 0.15960544479243896 - MAE: 0.3032567813495726
Epoch: 6
Training loss: 0.15132688611745834 - MAE: 0.29362002153133276
Validation loss : 0.15873736844343297 - MAE: 0.3026340556567635
Epoch: 7
Training loss: 0.15053230985999108 - MAE: 0.2936411939181263
Validation loss : 0.15721423354218989 - MAE: 0.3007017446245457
Epoch: 8
Training loss: 0.15200649231672286 - MAE: 0.29530517620818714
Validation loss : 0.1567032946383252 - MAE: 0.29992379573698985
Epoch: 9
Training loss: 0.14823424234986304 - MAE: 0.29201854833537316
Validation loss : 0.15661324560642242 - MAE: 0.3000581370785886
Epoch: 10
Training loss: 0.1488597995787859 - MAE: 0.29202925516443873
Validation loss : 0.15594959259033203 - MAE: 0.29955980657011466
Epoch: 11
Training loss: 0.1478568846732378 - MAE: 0.2920750352468466
Validation loss : 0.15519277882926605 - MAE: 0.29875912341148037
Epoch: 12
Training loss: 0.14643058851361274 - MAE: 0.2900920148543982
Validation loss : 0.1542835871086401 - MAE: 0.29761316691536144
Epoch: 13
Training loss: 0.145680733025074 - MAE: 0.29051563041966255
Validation loss : 0.1540197721299003 - MAE: 0.2970695665813526
Epoch: 14
Training loss: 0.14762464702129363 - MAE: 0.2914787752307884
Validation loss : 0.15417058765888214 - MAE: 0.2980900478295141
Epoch: 15
Training loss: 0.14693237617611885 - MAE: 0.2902377114756851
Validation loss : 0.15259952054304235 - MAE: 0.2949612107951529
Epoch: 16
Training loss: 0.14457482241094113 - MAE: 0.2876271989725693
Validation loss : 0.15327997593318715 - MAE: 0.2970857583493576
Epoch: 17
Training loss: 0.14548898428678514 - MAE: 0.28977346291513006
Validation loss : 0.15336684838813894 - MAE: 0.29746389708445636