File size: 3,030 Bytes
5552b94
ae2fd9c
a175971
da2905e
4e0388f
7ee0095
1032667
68e9b01
4d8ecee
d24808e
0e9b655
91cd278
1674edc
dccfaf7
b9ad253
667a823
7796f67
5cff094
cbf1e6b
b5f1c41
c034201
cd862b5
176932e
c5ce206
4e27620
9d367dc
1364cc1
b55770b
a713748
cfe319b
192d6f2
dc1c95e
75a8cc8
eda653b
d3d602f
ddf2d92
73f6309
60bafca
9afeddf
30452f0
3527ea1
41502e2
79d49ef
e92f396
56c80eb
47115c7
14599ab
2a07878
42b860c
b2a9c86
9c6556d
e03f879
f84dcfb
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
Started at: 11:40:24
nb-bert-base, 0.001, 512
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 0.6829635271659265 - MAE: 0.6693431870059429
Validation loss : 0.5015756666660309 - MAE: 0.5997545175999092
Epoch: 1
Training loss: 0.26068564561697155 - MAE: 0.40026651290975057
Validation loss : 0.2343240737915039 - MAE: 0.3639362216460188
Epoch: 2
Training loss: 0.20825687165443713 - MAE: 0.34524354469083296
Validation loss : 0.20746740698814392 - MAE: 0.35860255150045284
Epoch: 3
Training loss: 0.1804403788768328 - MAE: 0.324681643823719
Validation loss : 0.17072068750858307 - MAE: 0.3124492505853874
Epoch: 4
Training loss: 0.17333761086830726 - MAE: 0.3132351201497551
Validation loss : 0.17095611095428467 - MAE: 0.31916315360789926
Epoch: 5
Training loss: 0.16724205017089844 - MAE: 0.3070826757836269
Validation loss : 0.16323553621768952 - MAE: 0.30865708568301353
Epoch: 6
Training loss: 0.16199684257690722 - MAE: 0.3026997098344876
Validation loss : 0.1614771842956543 - MAE: 0.3067986414590729
Epoch: 7
Training loss: 0.16401460537543663 - MAE: 0.3037438752420999
Validation loss : 0.15994466841220856 - MAE: 0.30503234812758634
Epoch: 8
Training loss: 0.16340345602769119 - MAE: 0.30281380735994723
Validation loss : 0.1587683379650116 - MAE: 0.3037075319625505
Epoch: 9
Training loss: 0.16126641172629136 - MAE: 0.3015211312899261
Validation loss : 0.15727587938308715 - MAE: 0.30159571389132384
Epoch: 10
Training loss: 0.16030011841884026 - MAE: 0.29977143854425936
Validation loss : 0.1565701872110367 - MAE: 0.30090318825320855
Epoch: 11
Training loss: 0.15990765392780304 - MAE: 0.3001817094076697
Validation loss : 0.15565772354602814 - MAE: 0.29973778065557
Epoch: 12
Training loss: 0.15747033861967233 - MAE: 0.2969911949201895
Validation loss : 0.15501231849193572 - MAE: 0.2991009715387324
Epoch: 13
Training loss: 0.15866013902884263 - MAE: 0.2974925745491436
Validation loss : 0.15455043613910674 - MAE: 0.298642265264826
Epoch: 14
Training loss: 0.15624549755683312 - MAE: 0.29534914888328495
Validation loss : 0.15433281362056733 - MAE: 0.2986108357413379
Epoch: 15
Training loss: 0.15603324541678795 - MAE: 0.2951994368480578
Validation loss : 0.1534305989742279 - MAE: 0.2972280074655698
Epoch: 16
Training loss: 0.15464195150595444 - MAE: 0.2940797652071446