File size: 3,198 Bytes
b209950
fbbf5ad
45f1b11
8f2a114
5c745af
76f09e8
a43ad13
b6bf34a
4b521a4
c9d0e09
bf1909d
855c551
d6b12f1
39a6869
b135330
07c54a2
68e78a0
1362abc
3170096
cf591ec
a8c2817
0882573
702e361
342d348
caddab9
076cb72
15a9813
4fe6668
a3eea35
b32edbf
d81668b
9867f74
8ebb769
8f70b79
2e9d3fb
f6fe1e4
3367e7e
8816aa7
6edf92f
1383916
dc7fce0
7b508ea
9101793
9a95b4c
523d0cb
1f8f95f
f435dbf
ece12e4
d23c522
3eed329
53ee702
b14cf47
ab2d562
f4d7a58
6a27fda
7879ee0
c3a2894
ea83bc6
380ea57
568c2af
8c9976d
55b6c55
d953e35
8287bc7
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
Started at: 14:55:50
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.2707270306348801 - MAE: 0.4014548746281838
Validation loss : 0.18827014460283167 - MAE: 0.3350395914193632
Epoch: 1
Training loss: 0.18397963613271714 - MAE: 0.3274942520732981
Validation loss : 0.17896055035731373 - MAE: 0.32625804300542655
Epoch: 2
Training loss: 0.17076618894934653 - MAE: 0.3156013590381467
Validation loss : 0.17423390377970302 - MAE: 0.32323377271249604
Epoch: 3
Training loss: 0.1643083082139492 - MAE: 0.3095641840936019
Validation loss : 0.16728906508754282 - MAE: 0.3149824083534162
Epoch: 4
Training loss: 0.15957146167755126 - MAE: 0.3037565492992936
Validation loss : 0.16289721429347992 - MAE: 0.3089864069106985
Epoch: 5
Training loss: 0.15854436449706555 - MAE: 0.301500893432691
Validation loss : 0.16129758033682318 - MAE: 0.308202317708313
Epoch: 6
Training loss: 0.15570181787014006 - MAE: 0.29918709351469924
Validation loss : 0.16344413292758606 - MAE: 0.30938825983369334
Epoch: 7
Training loss: 0.15292704798281193 - MAE: 0.2961117975066867
Validation loss : 0.16081692103077383 - MAE: 0.3033625257590202
Epoch: 8
Training loss: 0.15216302141547203 - MAE: 0.2959306926556599
Validation loss : 0.15965510378865636 - MAE: 0.30640949969727455
Epoch: 9
Training loss: 0.1456899941712618 - MAE: 0.2888247514317626
Validation loss : 0.15742756864603827 - MAE: 0.3014585494849406
Epoch: 10
Training loss: 0.1467710939049721 - MAE: 0.2907957975079187
Validation loss : 0.15912020206451416 - MAE: 0.30151619716495687
Epoch: 11
Training loss: 0.14659826673567294 - MAE: 0.2893981534314371
Validation loss : 0.1602539621731814 - MAE: 0.3022486893374817
Epoch: 12
Training loss: 0.14781650044023992 - MAE: 0.29275159014985785
Validation loss : 0.15743818177896388 - MAE: 0.3011467784998341
Epoch: 13
Training loss: 0.14709322392940521 - MAE: 0.2905235164202929
Validation loss : 0.15774143706349766 - MAE: 0.30035432342912505
Epoch: 14
Training loss: 0.14297918483614921 - MAE: 0.2870641656789985
Validation loss : 0.16193263846285202 - MAE: 0.3045518551046045
Epoch: 15
Training loss: 0.14410636112093925 - MAE: 0.2889387564652954
Validation loss : 0.15722107492825566 - MAE: 0.299146052737687
Epoch: 16
Training loss: 0.14345928631722926 - MAE: 0.28585450120342
Validation loss : 0.15835655086180744 - MAE: 0.2999799426627903
Epoch: 17
Training loss: 0.14413826659321785 - MAE: 0.2881013153573912
Validation loss : 0.15604389984818065 - MAE: 0.29791508329064814
Epoch: 18
Training loss: 0.14309908427298068 - MAE: 0.2863054251735944
Validation loss : 0.15559008366921367 - MAE: 0.2975231734983981
Epoch: 19
Training loss: 0.14340508081018924 - MAE: 0.2865255988742361
Validation loss : 0.15817041651291006 - MAE: 0.30007397316689965
Epoch: 20
Training loss: 0.14153706684708595 - MAE: 0.28535710740880754