File size: 2,837 Bytes
bbb722c
4c157ac
2bde72e
3081378
16c7808
3cc8443
9d9ea54
debea42
4bd8d4e
25b4faa
a29f512
36bda71
745b97f
2f3d716
2cb0d6f
eb83d32
5aeb413
686fa7c
45248f0
4a5b452
c2d5a7f
bb137ee
8fee9fc
c6b6989
41fde32
f7a96cd
3c7fcf3
b8dd206
15fc6d8
9cb438d
10c488e
20e2cf6
94c9d98
4e97b3f
87790cb
c915edc
a1a8ba3
d8d14d9
994bffd
2414614
a6bb4cc
dfa8f17
13835ec
5f49bea
820f0eb
e35aeaf
33393bd
f0a5ed6
378a5b0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
Started at: 13:11:04
nb-bert-base, 0.001, 256
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 0.47113645792007447 - MAE: 0.5372690642533006
Validation loss : 0.25360237227545845 - MAE: 0.38100557323888096
Epoch: 1
Training loss: 0.1951601755619049 - MAE: 0.3381085981919985
Validation loss : 0.16845231420463985 - MAE: 0.3146699028671109
Epoch: 2
Training loss: 0.1700725769996643 - MAE: 0.3132894074707096
Validation loss : 0.16252560251288944 - MAE: 0.3101315378266
Epoch: 3
Training loss: 0.16299866735935212 - MAE: 0.30571749229812295
Validation loss : 0.1575031735830837 - MAE: 0.3036086169337241
Epoch: 4
Training loss: 0.16002290964126586 - MAE: 0.3030143814677838
Validation loss : 0.1555153015587065 - MAE: 0.30161708462789844
Epoch: 5
Training loss: 0.15610405176877976 - MAE: 0.2988610182826646
Validation loss : 0.15254790584246317 - MAE: 0.2976352880446658
Epoch: 6
Training loss: 0.15590642750263214 - MAE: 0.2972598594587939
Validation loss : 0.15147964821921456 - MAE: 0.29633871076747587
Epoch: 7
Training loss: 0.15364289730787278 - MAE: 0.2946944064180254
Validation loss : 0.1510375224881702 - MAE: 0.29637298396301465
Epoch: 8
Training loss: 0.1531897461414337 - MAE: 0.29468431902613157
Validation loss : 0.1495788519581159 - MAE: 0.29459826933056066
Epoch: 9
Training loss: 0.15212808579206466 - MAE: 0.2927512017766453
Validation loss : 0.14972183108329773 - MAE: 0.2947980987493768
Epoch: 10
Training loss: 0.15067319005727767 - MAE: 0.29223722020557824
Validation loss : 0.14739697674910227 - MAE: 0.29118543288491733
Epoch: 11
Training loss: 0.14746269524097444 - MAE: 0.28925764259432013
Validation loss : 0.14714962244033813 - MAE: 0.2916633442698583
Epoch: 12
Training loss: 0.1482395714521408 - MAE: 0.29003915570395034
Validation loss : 0.14708403746287027 - MAE: 0.2910518509892996
Epoch: 13
Training loss: 0.1483614018559456 - MAE: 0.29189581430634054
Validation loss : 0.14528870417012107 - MAE: 0.28829864578681347
Epoch: 14
Training loss: 0.14468369722366334 - MAE: 0.2842549207700623
Validation loss : 0.14488287270069122 - MAE: 0.28816803717811224
Epoch: 15