File size: 2,739 Bytes
c20f414
b2fc17b
01c12e3
2ecef43
396ab70
56862d9
d72bc25
6d53e85
ee873ff
1030e04
db424fe
089be7b
b3f91d8
4d1b736
c3d23da
6d355ea
7b31ad3
51fab6f
daf9fa5
9132459
add69de
af5b932
fe040a9
6d84d6a
04fb77f
45d9452
8cc1242
fe42d1f
8599816
a62e89a
d200d6e
1035181
8845b3f
8e8bfbb
5900d89
5830f93
e9ebafd
c2ca7e3
279704b
5163a3d
19938fd
cd604b9
e0ed1d1
17ec8ee
daa5c03
ca5c403
029678b
97bee02
4a4b6c9
1aa6200
dc81459
cbb6fd8
4ad3b58
0ee94a7
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
Started at: 12:07:49
norbert2, 1e-06, 256
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.30783953905105593 - MAE: 0.43719580843584505
Validation loss : 0.21560138960679373 - MAE: 0.3596361274665927
Epoch: 1
Training loss: 0.2266489964723587 - MAE: 0.3668055945655404
Validation loss : 0.2019458363453547 - MAE: 0.34808209428471554
Epoch: 2
Training loss: 0.19879530966281891 - MAE: 0.3426665569097344
Validation loss : 0.18689965373939937 - MAE: 0.3323873764788491
Epoch: 3
Training loss: 0.19044960021972657 - MAE: 0.3341355262436057
Validation loss : 0.17861351039674547 - MAE: 0.32349605653410407
Epoch: 4
Training loss: 0.18339800894260405 - MAE: 0.3276666275734149
Validation loss : 0.17232751680745018 - MAE: 0.316849883292955
Epoch: 5
Training loss: 0.17674655199050904 - MAE: 0.32018679623163826
Validation loss : 0.16676822139157188 - MAE: 0.31070956479175915
Epoch: 6
Training loss: 0.17175395369529725 - MAE: 0.3132561194660417
Validation loss : 0.16173937420050302 - MAE: 0.30571354481973123
Epoch: 7
Training loss: 0.16629139453172684 - MAE: 0.3090430212290697
Validation loss : 0.1575402236647076 - MAE: 0.30089353544828384
Epoch: 8
Training loss: 0.16224935233592988 - MAE: 0.3062369513645003
Validation loss : 0.15363233288129172 - MAE: 0.29686793533262296
Epoch: 9
Training loss: 0.15677126228809357 - MAE: 0.3001369819741856
Validation loss : 0.1501677847570843 - MAE: 0.2930093392614317
Epoch: 10
Training loss: 0.1522154378890991 - MAE: 0.2944758665139634
Validation loss : 0.1469805646273825 - MAE: 0.2896878885467087
Epoch: 11
Training loss: 0.14997684746980666 - MAE: 0.2930107961544775
Validation loss : 0.14408415887090895 - MAE: 0.28684108941263253
Epoch: 12
Training loss: 0.1469377765059471 - MAE: 0.29026566478408594
Validation loss : 0.14135516103770998 - MAE: 0.2838060469660866
Epoch: 13
Training loss: 0.1430877184867859 - MAE: 0.2856650031673201
Validation loss : 0.1389518520898289 - MAE: 0.2805393504731374
Epoch: 14
Training loss: 0.14132562696933745 - MAE: 0.2830484027164898
Validation loss : 0.1361759594745106 - MAE: 0.27795312559437724
Epoch: 15
Training loss: 0.1375071695446968 - MAE: 0.2793070474640609
Validation loss : 0.1337506083978547 - MAE: 0.2752909288563907
Epoch: 16
Training loss: 0.13490652769804 - MAE: 0.27698289906517004
Validation loss : 0.13174073066976336 - MAE: 0.27296385689056446