File size: 3,181 Bytes
a8ac709
3859c25
6529817
4514b14
6e0a8f6
14f529a
a72986b
1a69ea7
f297e0c
442fc24
f4c2df2
5317f9f
c31f45e
5dc0118
78abf23
84557e7
2a1a3e9
a5e300e
95ea73f
c7c2512
cba1cef
5fb793b
bf159f7
863e0fc
e7b94b7
a9a1d34
ca6032c
1723a4c
0557470
00b4de5
10d55ad
c5ee41a
25bcbbf
c0aab7a
30791a3
95fb2c8
3ffe382
4c3228c
6e11297
82848a1
f76b568
ec0a5da
e3c0791
6b38722
598194b
e8ff732
b99b46e
cae74a4
5acb508
884518f
8eeaa80
ba9a760
a8f29e0
aef72bc
83dc6f3
2968a05
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
Started at: 12:25:48
nb-bert-base, 5e-06, 256
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 0.6139226149428975 - MAE: 0.6256533578877125
Validation loss : 0.38856891581886693 - MAE: 0.5055659858766396
Epoch: 1
Training loss: 0.35679552988572555 - MAE: 0.47769355546848036
Validation loss : 0.2988674891622443 - MAE: 0.427160547565342
Epoch: 2
Training loss: 0.2514720949259671 - MAE: 0.3919834602387325
Validation loss : 0.20192746033794 - MAE: 0.35617145796178046
Epoch: 3
Training loss: 0.18519783670252019 - MAE: 0.34110099718701625
Validation loss : 0.16307617410233147 - MAE: 0.31849970460916255
Epoch: 4
Training loss: 0.15608852465044368 - MAE: 0.3137766481088091
Validation loss : 0.14197467973357752 - MAE: 0.29514997623851247
Epoch: 5
Training loss: 0.13567890524864196 - MAE: 0.29194648927171835
Validation loss : 0.12267970411401045 - MAE: 0.27265562083225603
Epoch: 6
Training loss: 0.11870894933288748 - MAE: 0.2718382958457312
Validation loss : 0.11097796692659981 - MAE: 0.25758373825453007
Epoch: 7
Training loss: 0.10573411136865615 - MAE: 0.25600807072103077
Validation loss : 0.1004621064976642 - MAE: 0.24238915549512713
Epoch: 8
Training loss: 0.095039548386227 - MAE: 0.24105853745370692
Validation loss : 0.08920641753234361 - MAE: 0.22744885245923066
Epoch: 9
Training loss: 0.08564553965221752 - MAE: 0.2290994268022599
Validation loss : 0.08278850426799372 - MAE: 0.21588505087327514
Epoch: 10
Training loss: 0.07813239084048705 - MAE: 0.21902914877383883
Validation loss : 0.07834271557236973 - MAE: 0.20735584144221048
Epoch: 11
Training loss: 0.07202434621073983 - MAE: 0.20906077481733826
Validation loss : 0.07437812458527715 - MAE: 0.20084316074607808
Epoch: 12
Training loss: 0.06499892032959244 - MAE: 0.19809261164608147
Validation loss : 0.07147281244397163 - MAE: 0.19576095041947425
Epoch: 13
Training loss: 0.05931831347671422 - MAE: 0.1885207176175341
Validation loss : 0.06822822988033295 - MAE: 0.18961870826028115
Epoch: 14
Training loss: 0.05735694515434178 - MAE: 0.1851404494180537
Validation loss : 0.07027043755117215 - MAE: 0.1932250628961616
Epoch: 15
Training loss: 0.05395592573014173 - MAE: 0.17943802032598533
Validation loss : 0.06585650106793955 - MAE: 0.18582157106086558
Epoch: 16
Training loss: 0.049961398745124994 - MAE: 0.1721245932478966
Validation loss : 0.06163646065090832 - MAE: 0.17768078284064118
Epoch: 17
Training loss: 0.04718107994307171 - MAE: 0.16805412913590112