File size: 5,361 Bytes
841080a
53ac994
d876328
a667809
b252260
36060c1
b970158
dfb5123
9ab82d4
e544147
af506e4
cad50a0
990e359
0965f2d
5232ecc
162f5ae
9bd8049
1245df1
40db2a9
f1861e8
ca66ff9
90a84d3
a0e5fd9
7753880
76cd010
6988c4e
da01ac6
d5d07e2
17fa806
bd661cc
a75bb70
1a76d41
b130ddf
12bb8d9
645b211
2715ef8
4418613
3c51552
acbbf33
fdc846d
c5d66e9
79fd51d
cd09b45
7cb372e
93d3961
7921e72
9bd6a81
e18d985
afb3eed
012a19d
0a3911e
80f044a
20a81b2
cd6e805
10888ce
0caf18f
46a2c3e
aa8b0de
1e5ea82
bc05418
18b0294
eef8046
a0547fe
a22e795
1476aa3
efae211
b6fe52b
fdd38aa
62e595e
d129992
05e35d6
7135bf6
c87ee6b
2c86dfc
bcadaae
06c7ca0
a63fe2f
cb2a9fa
f8a6bff
fd2e4f1
dd00538
9de94e7
72a0310
897bef8
dc3df3b
757aabf
8811f92
555e0fd
fd5c3cb
c395f7c
a2d5860
ca65a32
44f4c55
2ef0bfb
19eac4f
0cd6ec1
7f40813
ca7bb25
ea17e7c
44ebf93
b5bb883
8238019
32c7b3a
3a2942b
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
Started at: 15:31:13
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 1.2954848909378052 - MAE: 1.050244734015666
Validation loss : 0.8470989664395651 - MAE: 0.8281701794455048
Epoch: 1
Training loss: 0.4445240414142609 - MAE: 0.5411769913687082
Validation loss : 0.324721054898368 - MAE: 0.43869614013342645
Epoch: 2
Training loss: 0.25153143346309664 - MAE: 0.3816801249414825
Validation loss : 0.23877331614494324 - MAE: 0.3700370086786411
Epoch: 3
Training loss: 0.23011713743209838 - MAE: 0.36750653866750815
Validation loss : 0.22173692286014557 - MAE: 0.3575197016352192
Epoch: 4
Training loss: 0.21643910229206084 - MAE: 0.35473120380428824
Validation loss : 0.2111869735850228 - MAE: 0.3495476388776553
Epoch: 5
Training loss: 0.20841679513454436 - MAE: 0.3475277955848527
Validation loss : 0.2027071151468489 - MAE: 0.3434146884806285
Epoch: 6
Training loss: 0.20075951278209686 - MAE: 0.3414739466792711
Validation loss : 0.19626588291592068 - MAE: 0.3390224698180928
Epoch: 7
Training loss: 0.1959322625398636 - MAE: 0.33718433600506764
Validation loss : 0.19155074821578133 - MAE: 0.33601220203771287
Epoch: 8
Training loss: 0.19064979493618012 - MAE: 0.33252034985072276
Validation loss : 0.18807757563061184 - MAE: 0.33369018106396875
Epoch: 9
Training loss: 0.18758945107460023 - MAE: 0.3310842562717609
Validation loss : 0.18544620275497437 - MAE: 0.33179714328905824
Epoch: 10
Training loss: 0.18699397385120392 - MAE: 0.3293496073915552
Validation loss : 0.1833840227789349 - MAE: 0.3305601596359932
Epoch: 11
Training loss: 0.1838909602165222 - MAE: 0.326863464187819
Validation loss : 0.18136033746931288 - MAE: 0.32891395485147645
Epoch: 12
Training loss: 0.18286386430263518 - MAE: 0.325863773872513
Validation loss : 0.17971531218952602 - MAE: 0.32756646233438397
Epoch: 13
Training loss: 0.1797239762544632 - MAE: 0.32309299652089063
Validation loss : 0.17807276712523568 - MAE: 0.32608809156272295
Epoch: 14
Training loss: 0.17857051372528077 - MAE: 0.3213261058182539
Validation loss : 0.1767602562904358 - MAE: 0.3250377233246472
Epoch: 15
Training loss: 0.17911633133888244 - MAE: 0.3217271718139446
Validation loss : 0.1751896556880739 - MAE: 0.3234474552955444
Epoch: 16
Training loss: 0.17600405514240264 - MAE: 0.3189804320611656
Validation loss : 0.17363134854369694 - MAE: 0.3217130704298945
Epoch: 17
Training loss: 0.17391868591308593 - MAE: 0.3165747054747144
Validation loss : 0.17256545523802438 - MAE: 0.3208692052839463
Epoch: 18
Training loss: 0.17313318729400634 - MAE: 0.3169905761419007
Validation loss : 0.17127088208993277 - MAE: 0.3194913697705306
Epoch: 19
Training loss: 0.17114326536655425 - MAE: 0.31336970082229104
Validation loss : 0.17016845610406664 - MAE: 0.31830314372899243
Epoch: 20
Training loss: 0.1727628666162491 - MAE: 0.31565258354376935
Validation loss : 0.16911302506923676 - MAE: 0.31717762732632004
Epoch: 21
Training loss: 0.17012356221675873 - MAE: 0.31316639736858143
Validation loss : 0.1677489462825987 - MAE: 0.31546195651667186
Epoch: 22
Training loss: 0.16859103083610535 - MAE: 0.3127057693660738
Validation loss : 0.16697113381491768 - MAE: 0.31471562544725196
Epoch: 23
Training loss: 0.1686873698234558 - MAE: 0.311427263715106
Validation loss : 0.16612334880563948 - MAE: 0.3138178950174241
Epoch: 24
Training loss: 0.16640773981809617 - MAE: 0.3095450918700528
Validation loss : 0.16508907741970485 - MAE: 0.3125503103864031
Epoch: 25
Training loss: 0.1658923715353012 - MAE: 0.3082243922152051
Validation loss : 0.16451590259869894 - MAE: 0.3120238609455365
Epoch: 26
Training loss: 0.16436728775501253 - MAE: 0.3071946399101465
Validation loss : 0.16374372442563376 - MAE: 0.3111470200747217
Epoch: 27
Training loss: 0.16651157915592193 - MAE: 0.30963943402408695
Validation loss : 0.16270800100432503 - MAE: 0.3097923430317113
Epoch: 28
Training loss: 0.16491072356700898 - MAE: 0.3070697425803366
Validation loss : 0.1623369190427992 - MAE: 0.3095072696428577
Epoch: 29
Training loss: 0.16389427542686463 - MAE: 0.30555235896137595
Validation loss : 0.16124575005637276 - MAE: 0.308045409739515
Epoch: 30
Training loss: 0.16372050642967223 - MAE: 0.3068781009000078
Validation loss : 0.1606789148516125 - MAE: 0.3074341210225871
Epoch: 31
Training loss: 0.16316969752311705 - MAE: 0.3060961831725536
Validation loss : 0.16020016206635368 - MAE: 0.3069771875888861
Epoch: 32
Training loss: 0.1627587217092514 - MAE: 0.30426592400661
Validation loss : 0.15954788691467708 - MAE: 0.3061654847578481
Epoch: 33
Training loss: 0.16183113038539887 - MAE: 0.30331386520448744
Validation loss : 0.15876509911484188 - MAE: 0.30512034015868056