File size: 2,362 Bytes
a8ac709
3859c25
6529817
4514b14
6e0a8f6
14f529a
a72986b
1a69ea7
f297e0c
442fc24
f4c2df2
5317f9f
c31f45e
5dc0118
78abf23
84557e7
2a1a3e9
a5e300e
95ea73f
c7c2512
cba1cef
5fb793b
bf159f7
863e0fc
e7b94b7
a9a1d34
ca6032c
1723a4c
0557470
00b4de5
10d55ad
c5ee41a
25bcbbf
c0aab7a
30791a3
95fb2c8
3ffe382
4c3228c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
Started at: 12:25:48
nb-bert-base, 5e-06, 256
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 0.6139226149428975 - MAE: 0.6256533578877125
Validation loss : 0.38856891581886693 - MAE: 0.5055659858766396
Epoch: 1
Training loss: 0.35679552988572555 - MAE: 0.47769355546848036
Validation loss : 0.2988674891622443 - MAE: 0.427160547565342
Epoch: 2
Training loss: 0.2514720949259671 - MAE: 0.3919834602387325
Validation loss : 0.20192746033794 - MAE: 0.35617145796178046
Epoch: 3
Training loss: 0.18519783670252019 - MAE: 0.34110099718701625
Validation loss : 0.16307617410233147 - MAE: 0.31849970460916255
Epoch: 4
Training loss: 0.15608852465044368 - MAE: 0.3137766481088091
Validation loss : 0.14197467973357752 - MAE: 0.29514997623851247
Epoch: 5
Training loss: 0.13567890524864196 - MAE: 0.29194648927171835
Validation loss : 0.12267970411401045 - MAE: 0.27265562083225603
Epoch: 6
Training loss: 0.11870894933288748 - MAE: 0.2718382958457312
Validation loss : 0.11097796692659981 - MAE: 0.25758373825453007
Epoch: 7
Training loss: 0.10573411136865615 - MAE: 0.25600807072103077
Validation loss : 0.1004621064976642 - MAE: 0.24238915549512713
Epoch: 8
Training loss: 0.095039548386227 - MAE: 0.24105853745370692
Validation loss : 0.08920641753234361 - MAE: 0.22744885245923066
Epoch: 9
Training loss: 0.08564553965221752 - MAE: 0.2290994268022599
Validation loss : 0.08278850426799372 - MAE: 0.21588505087327514
Epoch: 10
Training loss: 0.07813239084048705 - MAE: 0.21902914877383883
Validation loss : 0.07834271557236973 - MAE: 0.20735584144221048
Epoch: 11
Training loss: 0.07202434621073983 - MAE: 0.20906077481733826