File size: 2,769 Bytes
ad09416
88a1958
62e4c2c
d9596f3
8e17eae
0fe6e72
78dbb4f
d515e03
b91dbf4
6a2360d
56fb8aa
57ca7b4
6f14c22
929b2fd
b61e7f6
f0efbae
f51b257
9befc0e
5ffd261
6acb753
12b3bef
ba5ce74
ba9c8c7
38a4779
0ecf70e
c9b470c
62a72bb
7851d3f
98cbb31
8f6eed9
73998f1
c4069e2
2b98d0d
fd08ee2
56de0cf
cccdbb2
1a0205c
b9d2e27
9ca0b60
6a0f0d6
a7df22e
5563041
cf1df83
2035e53
a0256d6
56129fa
822601a
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
Started at: 13:09:04
nb-bert-base, 5e-06, 256
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 0.5433284169435502 - MAE: 0.5775941266642696
Validation loss : 0.2084534830517239 - MAE: 0.34436377863461687
Epoch: 1
Training loss: 0.18881268858909606 - MAE: 0.3319951213456936
Validation loss : 0.17533069021171993 - MAE: 0.31952199275705395
Epoch: 2
Training loss: 0.17236681401729584 - MAE: 0.31305812195355187
Validation loss : 0.16540057957172394 - MAE: 0.30917810361613507
Epoch: 3
Training loss: 0.16512651681900026 - MAE: 0.30562505926503336
Validation loss : 0.1583948234717051 - MAE: 0.2991497249751852
Epoch: 4
Training loss: 0.15748567909002303 - MAE: 0.2989958723461345
Validation loss : 0.15284320215384165 - MAE: 0.29026856200804096
Epoch: 5
Training loss: 0.14964943289756774 - MAE: 0.2896343934318654
Validation loss : 0.14675226559241614 - MAE: 0.2830523245210482
Epoch: 6
Training loss: 0.1409749785065651 - MAE: 0.2818433281871407
Validation loss : 0.14069517536295784 - MAE: 0.2762524538160992
Epoch: 7
Training loss: 0.1349916523694992 - MAE: 0.27349521608195476
Validation loss : 0.13396325210730234 - MAE: 0.26954649511104584
Epoch: 8
Training loss: 0.12729438930749892 - MAE: 0.2662118251014631
Validation loss : 0.12773916290866005 - MAE: 0.2634949845222456
Epoch: 9
Training loss: 0.1206126120686531 - MAE: 0.25962342232666624
Validation loss : 0.12221866349379222 - MAE: 0.2589877239039535
Epoch: 10
Training loss: 0.11287372469902039 - MAE: 0.2514929032844295
Validation loss : 0.12077544712358051 - MAE: 0.25974176291313755
Epoch: 11
Training loss: 0.10924365520477294 - MAE: 0.24930425533115355
Validation loss : 0.11802816722128126 - MAE: 0.2588219114054121
Epoch: 12
Training loss: 0.10474763780832291 - MAE: 0.24292320602673487
Validation loss : 0.1159117536412345 - MAE: 0.25783798245851147
Epoch: 13
Training loss: 0.10213902741670608 - MAE: 0.24076924506931538
Validation loss : 0.11515615466568205 - MAE: 0.2585900787578427
Epoch: 14
Training loss: 0.09848966777324676 - MAE: 0.23697548065117832