File size: 2,923 Bytes
1f0e714
feee2fb
afd0df6
cedc360
4076651
198bc82
9be681b
635718c
6f37545
f4868cc
af83f0f
01d8801
3821530
c7b074c
918963b
b018bd1
63ab209
c5567dd
a71b618
a629578
1010e1d
c0de51a
e561a10
59315ef
6f2a08e
f5d7e65
35ad807
f965038
3976f6f
8ef0909
8db24bc
90c7d14
df6f1ec
2878461
e339fae
ed57cb7
1f77e8b
753cd9d
eccf51c
5b9357d
db04a2e
9e3ca02
299f942
bf4c388
cee2e35
3d23b7a
0b6872d
3e0f339
84dc40b
7130898
6dd7293
1f74be9
06651c4
04b59e7
3fae253
6a92ac2
cbf53a8
dfac188
579bd0e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
Started at: 21:52:14
norbert2, 0.001, 64
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.2395064114320158 - MAE: 0.3744385315534662
Validation loss : 0.20704752658352707 - MAE: 0.34597984502789736
Epoch: 1
Training loss: 0.1767307206085234 - MAE: 0.3192073445602084
Validation loss : 0.17625425801132666 - MAE: 0.3170721942019411
Epoch: 2
Training loss: 0.16618126668412275 - MAE: 0.3092801725863174
Validation loss : 0.17076683383096347 - MAE: 0.3112472933223099
Epoch: 3
Training loss: 0.16126501003299096 - MAE: 0.30584279932575203
Validation loss : 0.16347124337246924 - MAE: 0.3030033733379743
Epoch: 4
Training loss: 0.15625435588034717 - MAE: 0.29963192745208655
Validation loss : 0.15993442521853882 - MAE: 0.3004842166716503
Epoch: 5
Training loss: 0.15695305243887084 - MAE: 0.3010549775416317
Validation loss : 0.16180144634210702 - MAE: 0.3024127057273353
Epoch: 6
Training loss: 0.15556791018355975 - MAE: 0.29878062763401153
Validation loss : 0.15521211989901282 - MAE: 0.2970759091608139
Epoch: 7
Training loss: 0.15237431160428308 - MAE: 0.2953012408610363
Validation loss : 0.1589137056108677 - MAE: 0.29914319058997385
Epoch: 8
Training loss: 0.15153679695695338 - MAE: 0.29506802949496486
Validation loss : 0.15345649150284854 - MAE: 0.29414442352803954
Epoch: 9
Training loss: 0.1525212698663124 - MAE: 0.2963382092044523
Validation loss : 0.1529899320819161 - MAE: 0.29337032979089
Epoch: 10
Training loss: 0.15221869283252293 - MAE: 0.2953835321481793
Validation loss : 0.15726926471247818 - MAE: 0.29654991958475474
Epoch: 11
Training loss: 0.15084269725614124 - MAE: 0.29360040089351036
Validation loss : 0.15456406478628967 - MAE: 0.2942018313412496
Epoch: 12
Training loss: 0.14898980709940496 - MAE: 0.292388587665275
Validation loss : 0.15269274806434457 - MAE: 0.292333423595863
Epoch: 13
Training loss: 0.14591557916366693 - MAE: 0.2891725097986123
Validation loss : 0.1531008021398024 - MAE: 0.29286502452232843
Epoch: 14
Training loss: 0.15047907114329964 - MAE: 0.2928827053558407
Validation loss : 0.15025645133220789 - MAE: 0.2906735936090682
Epoch: 15
Training loss: 0.15035037162026973 - MAE: 0.2942172721654035
Validation loss : 0.1525195986032486 - MAE: 0.2924193384927183
Epoch: 16
Training loss: 0.14917351215174704 - MAE: 0.2934477894984301
Validation loss : 0.15107204539306235 - MAE: 0.2908045947936088
Epoch: 17
Training loss: 0.1479782590059319 - MAE: 0.2918790273580605
Validation loss : 0.15280635677503818 - MAE: 0.29140434010156746
Stopped after 17 epochs.
Prediction MAE: 0.2771