File size: 2,879 Bytes
c2db821
40a27d0
68f4497
1f4a22c
b4ffd13
d5a4695
c6e6584
dc5d88b
dd52d48
b61f21a
09d3954
090fd41
6db40fb
54cf612
e7553c2
2945aa5
aaa4b44
2fbfd07
8317af0
5869bac
a28638f
336628c
e268e1a
7d06ba6
56daa8b
31465fb
099d82c
caf2bfc
b79dae4
1017a33
69594a6
951bc42
ff5768f
3a7b877
12ce733
86f5fe0
3f4269a
5fae403
afb6c0a
a7357d8
94febcd
aa754d8
d3d65b1
2ea56ff
41742e2
35da941
4533262
b6f0c57
6f25c96
1316ad1
ff9253c
2b6167c
f7715ce
f475880
2fb79ff
379bea7
0c2fc47
0f78796
6d7013e
d287a8c
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
Started at: 23:31:07
norbert2, 0.001, 128
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.26663439720869064 - MAE: 0.3988883059100627
Validation loss : 0.18966426481218898 - MAE: 0.3369448800337227
Epoch: 1
Training loss: 0.1841131156682968 - MAE: 0.3278432048597948
Validation loss : 0.17534208385383382 - MAE: 0.3222601691590387
Epoch: 2
Training loss: 0.16919755205512046 - MAE: 0.3150248810255783
Validation loss : 0.17205075919628143 - MAE: 0.32034478399614935
Epoch: 3
Training loss: 0.16370050594210625 - MAE: 0.30848461858803294
Validation loss : 0.16791229563600876 - MAE: 0.3158996427097044
Epoch: 4
Training loss: 0.15879319690167903 - MAE: 0.30243582822506326
Validation loss : 0.1651218849069932 - MAE: 0.3108875101513454
Epoch: 5
Training loss: 0.15427802219986916 - MAE: 0.29842854069426583
Validation loss : 0.16311591600670533 - MAE: 0.3086705140377803
Epoch: 6
Training loss: 0.1524083361774683 - MAE: 0.29523006209175195
Validation loss : 0.163714456207612 - MAE: 0.30915422006445037
Epoch: 7
Training loss: 0.15138228960335254 - MAE: 0.29511583224000026
Validation loss : 0.16111911833286285 - MAE: 0.3054659731683512
Epoch: 8
Training loss: 0.1489515230804682 - MAE: 0.2933538312252247
Validation loss : 0.1581758834859904 - MAE: 0.3017375955992824
Epoch: 9
Training loss: 0.1511006211489439 - MAE: 0.29474934953223647
Validation loss : 0.16293136118089452 - MAE: 0.3091806150526381
Epoch: 10
Training loss: 0.14791207369416953 - MAE: 0.29105243780221857
Validation loss : 0.15650257950319962 - MAE: 0.3001227490893128
Epoch: 11
Training loss: 0.14760239347815513 - MAE: 0.2913725760078877
Validation loss : 0.16039451606133404 - MAE: 0.3021585680923023
Epoch: 12
Training loss: 0.1475458946824074 - MAE: 0.29206565534853324
Validation loss : 0.15759038968997843 - MAE: 0.2991883971577083
Epoch: 13
Training loss: 0.1458326653391123 - MAE: 0.28907803451268993
Validation loss : 0.155505466987105 - MAE: 0.29847763216785667
Epoch: 14
Training loss: 0.14564877450466157 - MAE: 0.29022519197923985
Validation loss : 0.16000595338204326 - MAE: 0.30122497450380015
Epoch: 15
Training loss: 0.14562662817537786 - MAE: 0.2891785655793424
Validation loss : 0.15955582874662735 - MAE: 0.30044914898196184
Epoch: 16
Training loss: 0.14269623432308434 - MAE: 0.2864237342804197
Validation loss : 0.16522828883984508 - MAE: 0.30640706274809065
Stopped after 16 epochs.
Prediction MAE: 0.2882
Finished at: 23:31:07
Time taken: 5698 s.
0 days 1 hours 34 minutes 58 seconds