File size: 3,068 Bytes
0292aec
2971ef0
2593aca
102eb3a
354d201
ffc7072
0eb949f
92ef2a4
ee5ab8c
beca0d6
8ba5556
c8f58b8
174ce95
5b89728
73c501e
ef1dd38
d5013d1
a135643
a121f85
7cb752a
d5b5d84
9ec22ba
864e68b
bcec72d
1969dbb
8cd73d3
212c5de
0bbbbd5
8a7ed8b
b6bf7ae
848b5ec
2c7cbb6
6ca1e82
af8694a
ca62c84
e05379d
0d22bc9
5140fd8
91c431c
b180168
953253d
d736cec
97df9a6
b95567b
aaecb39
1333621
1ea4569
37b68e8
2237eba
c45c5d2
0080abe
de9a60b
283583d
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
Started at: 11:40:26
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 0.4581766623258591 - MAE: 0.5292641480208709
Validation loss : 0.22879805664221445 - MAE: 0.3594399771230401
Epoch: 1
Training loss: 0.19436429917812348 - MAE: 0.3385698724059062
Validation loss : 0.17431378530131447 - MAE: 0.3219149815721847
Epoch: 2
Training loss: 0.17444411993026734 - MAE: 0.31684918185002736
Validation loss : 0.16497086816363865 - MAE: 0.31185002325092886
Epoch: 3
Training loss: 0.1655295819044113 - MAE: 0.3086558034500663
Validation loss : 0.15977457662423453 - MAE: 0.3062515869775394
Epoch: 4
Training loss: 0.16229379653930665 - MAE: 0.3037281934486638
Validation loss : 0.15585461341672474 - MAE: 0.3013561711510056
Epoch: 5
Training loss: 0.15735343039035798 - MAE: 0.299567204911211
Validation loss : 0.1539518212278684 - MAE: 0.2993376751781608
Epoch: 6
Training loss: 0.15725925147533418 - MAE: 0.2996774680828566
Validation loss : 0.15224697606431115 - MAE: 0.2976658126019026
Epoch: 7
Training loss: 0.15550604462623596 - MAE: 0.2963955747939702
Validation loss : 0.15032305154535505 - MAE: 0.29504313159460227
Epoch: 8
Training loss: 0.1532408055663109 - MAE: 0.2955262865803256
Validation loss : 0.1494835408197509 - MAE: 0.2939865938164665
Epoch: 9
Training loss: 0.15284976929426194 - MAE: 0.2941793210474116
Validation loss : 0.1482765343454149 - MAE: 0.2924292223271092
Epoch: 10
Training loss: 0.15122360587120057 - MAE: 0.29247309200113303
Validation loss : 0.14710667315456602 - MAE: 0.2907982624220039
Epoch: 11
Training loss: 0.14901623457670213 - MAE: 0.290842407098567
Validation loss : 0.14528998070293003 - MAE: 0.28806596351171554
Epoch: 12
Training loss: 0.15042486935853958 - MAE: 0.29147855511346055
Validation loss : 0.14535025589995915 - MAE: 0.2886085535457345
Epoch: 13
Training loss: 0.1494656953215599 - MAE: 0.2891299573328347
Validation loss : 0.14426815013090769 - MAE: 0.2873694180544381
Epoch: 14
Training loss: 0.1479952174425125 - MAE: 0.28811694403834054
Validation loss : 0.14319962759812674 - MAE: 0.28503047862306335
Epoch: 15
Training loss: 0.14743517249822616 - MAE: 0.2894441080912908
Validation loss : 0.1425529478324784 - MAE: 0.28427138674014546
Epoch: 16
Training loss: 0.1472056606411934 - MAE: 0.28826150001116063
Validation loss : 0.14231768581602308 - MAE: 0.2836957414049199