File size: 2,886 Bytes
521959b
c5c12d0
c26b107
f8e2c99
cf0f1d7
7ba0ab5
117d441
87139fd
51f7f0a
404af39
98d2b00
e30b1ec
ab10f85
c50c9c1
c596e21
b4bd98a
d0fa6c9
586cfd5
0323c9c
f0d45c2
19de30a
8ca0cf8
35f0f6c
a231784
fce29ad
cb9a13b
2713003
fef7c17
c3854d5
8ead488
a35b379
5c54f29
cf9b144
5a70457
8b0c28d
f712444
0aa527d
f00ed0b
2568793
011b9e5
41e0812
44a25c4
3ec0680
7ad050b
e25329d
f980ce7
2c0418c
ad931c4
e676cec
a8396d8
17d00bf
943bcd6
1752b11
d228e66
5131fb0
b02a646
b0c8671
569d1f2
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
Started at: 16:58:38
norbert, 1e-06, 128
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 0.7940866667032241 - MAE: 0.7335820706168809
Validation loss : 0.2009572711061029 - MAE: 0.347151339356108
Epoch: 1
Training loss: 0.19440142974257468 - MAE: 0.33696885945893024
Validation loss : 0.19513289893374725 - MAE: 0.34191775396532464
Epoch: 2
Training loss: 0.1880853456258774 - MAE: 0.32884479290870544
Validation loss : 0.1912411977263058 - MAE: 0.3377114550114297
Epoch: 3
Training loss: 0.18214607641100883 - MAE: 0.32429566931252046
Validation loss : 0.18406631753725164 - MAE: 0.33177290516437835
Epoch: 4
Training loss: 0.17814580470323563 - MAE: 0.3207397197424783
Validation loss : 0.1817777980776394 - MAE: 0.3288081865926719
Epoch: 5
Training loss: 0.17603084936738014 - MAE: 0.3188514731069929
Validation loss : 0.17798463649609508 - MAE: 0.3250048632803593
Epoch: 6
Training loss: 0.17104241520166397 - MAE: 0.3136062115504468
Validation loss : 0.1733261583482518 - MAE: 0.31932644642810926
Epoch: 7
Training loss: 0.16838639885187148 - MAE: 0.3123990239326607
Validation loss : 0.17248562416609595 - MAE: 0.31865962855570096
Epoch: 8
Training loss: 0.1654125490784645 - MAE: 0.3085900949825916
Validation loss : 0.16888807626331553 - MAE: 0.3153836119416533
Epoch: 9
Training loss: 0.1645794503390789 - MAE: 0.30737329828383714
Validation loss : 0.16706455805722406 - MAE: 0.3121367546740557
Epoch: 10
Training loss: 0.1616331295669079 - MAE: 0.30512118468554295
Validation loss : 0.16470523529192982 - MAE: 0.3095945078707051
Epoch: 11
Training loss: 0.16030726864933967 - MAE: 0.3040617280659885
Validation loss : 0.16272581894608104 - MAE: 0.3069817555674387
Epoch: 12
Training loss: 0.1589731414616108 - MAE: 0.3032846674015245
Validation loss : 0.16099156351650462 - MAE: 0.3047476777439796
Epoch: 13
Training loss: 0.15562866285443305 - MAE: 0.29989675715351916
Validation loss : 0.15904783939614014 - MAE: 0.3019063086532548
Epoch: 14
Training loss: 0.15696120083332063 - MAE: 0.30097456977168896
Validation loss : 0.15785902738571167 - MAE: 0.30075530215047686
Epoch: 15
Training loss: 0.151818914860487 - MAE: 0.2954353682482574
Validation loss : 0.1551657340982381 - MAE: 0.29802607762302763
Epoch: 16
Training loss: 0.1526854731142521 - MAE: 0.29552840146002957
Validation loss : 0.1538231990793172 - MAE: 0.29607686650726545
Epoch: 17
Training loss: 0.14886232271790503 - MAE: 0.2912935844145451
Validation loss : 0.15218366288086949 - MAE: 0.29439642167515684
Epoch: 18