File size: 3,249 Bytes
3d10b6c
15f5fdb
4b60a52
3daf40c
4d7c6a9
f8332a1
65c0083
259887a
7820e7c
acf4639
91b7d89
0d6d5cd
944333a
bf9120f
85c86a4
03b81a2
84369db
4a2af3a
f198f6e
df48e1e
274897a
21bdbdd
357a0eb
74df24e
1ffa540
c0f6cb7
7f41cf3
e142f0c
317e99d
24a1e00
142560f
782f080
d559341
dfdce49
9bc1cf4
88f3810
ad81937
c518c64
a730268
26b474f
c9fad4f
23c02e0
f5f0346
f6d3687
2d3c285
84b1da1
45b3cff
c883747
46a03e1
7772372
2c91b6c
c67f53d
428f0c5
4286e9a
033e9ce
2afcf9f
d741efe
f95737a
ac8a655
74d8fcf
0c56dfc
56eace8
2e9ac5e
3cca6ee
73ee24c
96df481
664da0f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
Started at: 07:38:00
norbert2, 0.001, 512
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.4007784449137174 - MAE: 0.5069355512324459
Validation loss : 0.3120305180549622 - MAE: 0.44266591438730646
Epoch: 1
Training loss: 0.2484535529063298 - MAE: 0.3928363123027111
Validation loss : 0.1947101354598999 - MAE: 0.3383142622305263
Epoch: 2
Training loss: 0.22325417743279383 - MAE: 0.36296588484431314
Validation loss : 0.182929190993309 - MAE: 0.3313696384637742
Epoch: 3
Training loss: 0.20250843579952532 - MAE: 0.3434091643314229
Validation loss : 0.18073179423809052 - MAE: 0.3283606648709959
Epoch: 4
Training loss: 0.19508753602321333 - MAE: 0.3348078961451888
Validation loss : 0.17457089722156524 - MAE: 0.32172754370186174
Epoch: 5
Training loss: 0.1839254005597188 - MAE: 0.32533722867728143
Validation loss : 0.171060973405838 - MAE: 0.317254907258527
Epoch: 6
Training loss: 0.18198314652993128 - MAE: 0.3254398824331828
Validation loss : 0.16855189800262452 - MAE: 0.31388433598330434
Epoch: 7
Training loss: 0.17925549011964065 - MAE: 0.32251319498782804
Validation loss : 0.16748179197311402 - MAE: 0.31228902521293
Epoch: 8
Training loss: 0.17594398214266851 - MAE: 0.3184498187707379
Validation loss : 0.16516335308551788 - MAE: 0.30966778780720217
Epoch: 9
Training loss: 0.1718841722378364 - MAE: 0.31540975077584577
Validation loss : 0.16371028423309325 - MAE: 0.3090980532981752
Epoch: 10
Training loss: 0.16989627939004165 - MAE: 0.3115659211880164
Validation loss : 0.16193810999393463 - MAE: 0.3079626352800581
Epoch: 11
Training loss: 0.1646776760999973 - MAE: 0.3064563203597883
Validation loss : 0.1605544239282608 - MAE: 0.30548879755410685
Epoch: 12
Training loss: 0.16483179307900941 - MAE: 0.3055755864052439
Validation loss : 0.15999217331409454 - MAE: 0.3059971345763114
Epoch: 13
Training loss: 0.16520174879294175 - MAE: 0.3064616799158078
Validation loss : 0.1594652861356735 - MAE: 0.30604126712541796
Epoch: 14
Training loss: 0.16016467259480402 - MAE: 0.30232788138904937
Validation loss : 0.1583849757909775 - MAE: 0.3044767570585575
Epoch: 15
Training loss: 0.15914533917720503 - MAE: 0.30153882059130266
Validation loss : 0.15729162096977234 - MAE: 0.30317250687226727
Epoch: 16
Training loss: 0.16054183932451102 - MAE: 0.3017137439411966
Validation loss : 0.15653180778026582 - MAE: 0.30260446283012854
Epoch: 17
Training loss: 0.15821081973039186 - MAE: 0.29985171444757386
Validation loss : 0.15718148052692413 - MAE: 0.30393464984074536
Epoch: 18
Training loss: 0.15764015110639426 - MAE: 0.29861296551168576
Validation loss : 0.15508275330066681 - MAE: 0.30125317290431686
Epoch: 19
Training loss: 0.15852965299899763 - MAE: 0.3010436145212502
Validation loss : 0.1547796756029129 - MAE: 0.3011046554502358
Prediction MAE: 0.2878
Finished at: 07:38:00
Time taken: 1488 s.
0 days 0 hours 24 minutes 48 seconds