File size: 3,531 Bytes
a9614e2
1c266c2
573d184
138c452
085fc0e
a8e4a32
60030c6
43aca8f
bced6f4
3c0ba3e
05cb146
2755fdc
6d43361
9ba6b1c
bc4b6fa
fd66402
8ac5557
0a5b886
7582378
7943a4b
1d207e4
fc0b9fb
778b2ee
d8f234d
b8f30b4
07d6b01
b1e9821
5d63be6
c240e23
7fd6a30
753d906
7988c17
f9cc4b4
2563dd5
3b32d9d
78c220f
13dd3c9
3645ffc
a67ad3c
1791d1d
417295b
9d01042
4692ce2
ee702fe
9ad9b31
8a55390
af3dc92
255d975
b5c59b2
6291ba6
910096e
f1cb193
5ecc42f
f52fd05
4aa4325
5b2f689
0ca188f
90fba87
3b6adf1
d7b4484
b3d9166
822020b
5f1eb05
5a03441
36bc79e
ff030dd
927b3c1
64f46f0
749183e
2c6fb37
bff4a2d
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
Started at: 21:34:05
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.32175946950912476 - MAE: 0.4437903503551826
Validation loss : 0.19621961481041378 - MAE: 0.3388283822410594
Epoch: 1
Training loss: 0.21072698533535003 - MAE: 0.35116530289867337
Validation loss : 0.1876767095592287 - MAE: 0.33628140968919695
Epoch: 2
Training loss: 0.1893942326307297 - MAE: 0.3312522516218666
Validation loss : 0.1748287214173211 - MAE: 0.3213797447248068
Epoch: 3
Training loss: 0.17969159841537474 - MAE: 0.3215382240664255
Validation loss : 0.1697687026527193 - MAE: 0.31446870224950263
Epoch: 4
Training loss: 0.17220519423484804 - MAE: 0.3129526796594085
Validation loss : 0.16811848680178323 - MAE: 0.3117262145282266
Epoch: 5
Training loss: 0.17296144247055054 - MAE: 0.31494873624825814
Validation loss : 0.16849000917540657 - MAE: 0.31104514179599985
Epoch: 6
Training loss: 0.16641035348176955 - MAE: 0.30768475494362546
Validation loss : 0.16421516074074638 - MAE: 0.30715220958263423
Epoch: 7
Training loss: 0.16057054400444032 - MAE: 0.30329494898002457
Validation loss : 0.161027698053254 - MAE: 0.304297376785587
Epoch: 8
Training loss: 0.1566170272231102 - MAE: 0.29957014870206655
Validation loss : 0.15982638630602095 - MAE: 0.30397747682822174
Epoch: 9
Training loss: 0.15673983812332154 - MAE: 0.3000540458404174
Validation loss : 0.159184659520785 - MAE: 0.3035235378542429
Epoch: 10
Training loss: 0.15630604147911073 - MAE: 0.29690365842430627
Validation loss : 0.15921704471111298 - MAE: 0.30240467396157955
Epoch: 11
Training loss: 0.15595020622015 - MAE: 0.29754135005638765
Validation loss : 0.15686986181471083 - MAE: 0.30119996351152656
Epoch: 12
Training loss: 0.15274528950452804 - MAE: 0.2944161972508913
Validation loss : 0.15652166141404045 - MAE: 0.3012713923501961
Epoch: 13
Training loss: 0.15205995708703995 - MAE: 0.29424324063629004
Validation loss : 0.1575678288936615 - MAE: 0.3004794443503115
Epoch: 14
Training loss: 0.15087180227041244 - MAE: 0.292504579327589
Validation loss : 0.15574459234873453 - MAE: 0.29932356111019714
Epoch: 15
Training loss: 0.1518820345401764 - MAE: 0.2940904971897683
Validation loss : 0.15581322544150883 - MAE: 0.30048684662514935
Epoch: 16
Training loss: 0.14916340589523316 - MAE: 0.2900369708605816
Validation loss : 0.15502946575482687 - MAE: 0.2986623058669649
Epoch: 17
Training loss: 0.14969733864068985 - MAE: 0.2916869417468108
Validation loss : 0.15631223718325296 - MAE: 0.2994015598567933
Epoch: 18
Training loss: 0.14621972769498826 - MAE: 0.2875086269286061
Validation loss : 0.1557358337773217 - MAE: 0.29931970436403404
Epoch: 19
Training loss: 0.14798570185899734 - MAE: 0.29024787778757843
Validation loss : 0.15424930387073094 - MAE: 0.2984811820958494
Epoch: 20
Training loss: 0.14769238144159316 - MAE: 0.2896109423923894
Validation loss : 0.1536422868569692 - MAE: 0.2987445844262027
Epoch: 21
Training loss: 0.14361368536949157 - MAE: 0.2842206176667335
Validation loss : 0.1533755792511834 - MAE: 0.29642898867488277
Epoch: 22
Training loss: 0.14456430345773696 - MAE: 0.2860957867005398
Validation loss : 0.1542035871081882 - MAE: 0.2971775973100257