File size: 3,227 Bytes
91debdd
98bf694
098a3a7
7d364d9
7de12fe
94e731c
11ca7ec
ab26d4b
168c6ee
09cf2e6
8b695f5
a49f76b
e48af9a
f520065
a01f89b
dc39c2d
27e06ed
2b6291f
47538e3
e58652e
a473844
15ffe81
77e0631
5a37c15
6f6abb9
125cd42
62a9067
05523d6
ed51ec3
41d68a1
f56a336
09fef35
d25394b
b539e21
43ef9c0
ff0b7a0
a25c8ea
fc4a173
19d8327
17b4ad0
68372e4
37c8ca5
05fc430
542f565
85183c9
97a6bbd
c3d1cc8
194c7d1
cbdd0dc
2f45103
f7a33fd
5344648
8117331
6080e06
549f056
904a782
f725763
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
Started at: 14:14:54
nb-bert-base, 1e-06, 256
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 1.3610501217842101 - MAE: 1.0817184245036757
Validation loss : 0.6300775276290046 - MAE: 0.6925418514553945
Epoch: 1
Training loss: 0.32747340738773345 - MAE: 0.4483029439383294
Validation loss : 0.24206096761756474 - MAE: 0.3825273214027049
Epoch: 2
Training loss: 0.22492031693458558 - MAE: 0.3634789782721722
Validation loss : 0.21290512879689535 - MAE: 0.3531350766820909
Epoch: 3
Training loss: 0.21074927151203154 - MAE: 0.34935127198342575
Validation loss : 0.19930076433552635 - MAE: 0.3423320980410668
Epoch: 4
Training loss: 0.1998182564973831 - MAE: 0.33952265508216506
Validation loss : 0.19053682684898376 - MAE: 0.33432257900743867
Epoch: 5
Training loss: 0.18897418558597565 - MAE: 0.3308955648791844
Validation loss : 0.183888531393475 - MAE: 0.32777869810289756
Epoch: 6
Training loss: 0.18244400441646577 - MAE: 0.3239016318289663
Validation loss : 0.17849575810962254 - MAE: 0.32227053870844763
Epoch: 7
Training loss: 0.17936479389667512 - MAE: 0.32048774343640357
Validation loss : 0.174004680580563 - MAE: 0.3172998359634526
Epoch: 8
Training loss: 0.17690573513507843 - MAE: 0.3179874754998065
Validation loss : 0.17019198503759173 - MAE: 0.3130601250011837
Epoch: 9
Training loss: 0.171479050219059 - MAE: 0.31243650862938
Validation loss : 0.16672976977295345 - MAE: 0.30925350280753894
Epoch: 10
Training loss: 0.16934633016586303 - MAE: 0.3105842075823956
Validation loss : 0.16365073952409956 - MAE: 0.3054956521646379
Epoch: 11
Training loss: 0.16901946306228638 - MAE: 0.3104853099071552
Validation loss : 0.16163094507323372 - MAE: 0.3018340222544562
Epoch: 12
Training loss: 0.16289644062519074 - MAE: 0.3032128571774679
Validation loss : 0.15861071149508157 - MAE: 0.29892911668430483
Epoch: 13
Training loss: 0.16094084948301315 - MAE: 0.3017364437829033
Validation loss : 0.15633813540140787 - MAE: 0.295844440886826
Epoch: 14
Training loss: 0.15619824528694154 - MAE: 0.2982343003784247
Validation loss : 0.15405048429965973 - MAE: 0.2929410397811803
Epoch: 15
Training loss: 0.15374296069145202 - MAE: 0.2942846683060764
Validation loss : 0.1519482500023312 - MAE: 0.2904616183074069
Epoch: 16
Training loss: 0.15236089646816253 - MAE: 0.29222253923753305
Validation loss : 0.15019716239637798 - MAE: 0.2879879020618735
Epoch: 17
Training loss: 0.15183452248573304 - MAE: 0.29124569910300246
Validation loss : 0.14797681156131956 - MAE: 0.2858146141464486