File size: 3,149 Bytes
1affc1c
18c0138
221e918
5272fa4
47622c4
d42f18b
d6b9d9e
6a061eb
4c7031e
fec7105
76fe34f
715d94a
78e089a
e2810c5
7120b98
3d73bfd
8475310
1ce0d0f
0432dbd
02e32ff
ac40506
d6157e1
9719423
221176c
6aa232b
1f1b310
dc07b22
f51ef5e
2594b9a
cd2b674
ee5ae6c
3d6bb3e
b5fb8d2
f425a3f
3b2e2e4
3842791
6b3ab6b
129a980
b768ab9
41920a2
bdb3a40
ce3c2a4
f3087db
372a28c
b4bbd81
b7362ab
a7c7a80
30933c7
1c92ebd
50adafe
b2716b6
59610ad
fbfae16
c639bcf
3d88cd4
d0f252d
53460ce
1ac6c6d
89ba1e6
31796f1
b216851
07055bf
b35743e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
Started at: 12:52:23
norbert, 5e-06, 256
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 0.6053906446153468 - MAE: 0.6184159507940801
Validation loss : 0.4188641560705085 - MAE: 0.5190200870427507
Epoch: 1
Training loss: 0.37205150289969013 - MAE: 0.4891842565138049
Validation loss : 0.3361840279478776 - MAE: 0.4554968620483818
Epoch: 2
Training loss: 0.30589349486611106 - MAE: 0.43993173479624126
Validation loss : 0.27830397376888677 - MAE: 0.41290173204153435
Epoch: 3
Training loss: 0.2584832324223085 - MAE: 0.4034832139365259
Validation loss : 0.2353723849120893 - MAE: 0.378263007542243
Epoch: 4
Training loss: 0.2208004041151567 - MAE: 0.3714158687728299
Validation loss : 0.20094563145386546 - MAE: 0.35106286126524094
Epoch: 5
Training loss: 0.1931754020127383 - MAE: 0.3488251679890293
Validation loss : 0.17955732188726725 - MAE: 0.33182885819892266
Epoch: 6
Training loss: 0.17624386982484297 - MAE: 0.3333650190493288
Validation loss : 0.16801361818062632 - MAE: 0.3229406276757107
Epoch: 7
Training loss: 0.16203992610627954 - MAE: 0.3199497565391834
Validation loss : 0.1510195426250759 - MAE: 0.306880161949399
Epoch: 8
Training loss: 0.15154187828302385 - MAE: 0.3089620812443599
Validation loss : 0.14486609556173025 - MAE: 0.29924708380840065
Epoch: 9
Training loss: 0.14079513807188382 - MAE: 0.29892621874422587
Validation loss : 0.13840533440050326 - MAE: 0.290965015323921
Epoch: 10
Training loss: 0.13332889432256873 - MAE: 0.29082796894473617
Validation loss : 0.12829295977165825 - MAE: 0.27816580614753955
Epoch: 11
Training loss: 0.12650601890954105 - MAE: 0.28307433471783455
Validation loss : 0.12042002497535002 - MAE: 0.26958139319431695
Epoch: 12
Training loss: 0.1180860929868438 - MAE: 0.27275313703304305
Validation loss : 0.1134116300626805 - MAE: 0.2581940768065212
Epoch: 13
Training loss: 0.1120438659732992 - MAE: 0.2657403258096752
Validation loss : 0.11130064725875854 - MAE: 0.2551686039616152
Epoch: 14
Training loss: 0.10740146054462953 - MAE: 0.26015119794445635
Validation loss : 0.10768120618242967 - MAE: 0.2487607443147535
Epoch: 15
Training loss: 0.10318632572889327 - MAE: 0.2533595643372031
Validation loss : 0.10223148920034107 - MAE: 0.2425870576871514
Epoch: 16
Training loss: 0.09851837293668227 - MAE: 0.2475603386691453
Validation loss : 0.10070653259754181 - MAE: 0.2398464623664194
Epoch: 17
Training loss: 0.09413933239199898 - MAE: 0.24082904543342942
Validation loss : 0.09791698738148338 - MAE: 0.23589278404154412
Epoch: 18
Training loss: 0.0905470837246288 - MAE: 0.23585713024943322
Validation loss : 0.09309150278568268 - MAE: 0.22871411303114594
Epoch: 19
Training loss: 0.08814722827889702 - MAE: 0.23258866689213922
Validation loss : 0.09581998304316872 - MAE: 0.23397831596428242