File size: 3,013 Bytes
d6c6bc9
3ddd5e8
91af41d
347bab1
ba4e8b8
1a31b88
89fb0a4
cbdc3a8
d22e7a3
89bc6e7
1748266
4a88d0c
ee484e2
5ad2f91
95bd9ae
71d7854
4dbfd5f
eba6ba2
bc43344
a2a3b80
617c641
44d7ae7
fb559d7
a91fb84
22a6fc6
51daeef
c509cc1
7196365
9997432
f516ba0
0c798c2
4b9cc99
1105b64
c4455be
47dd840
433ddfc
77433d7
d434c39
4d36c27
37b2133
ecc08b2
f9ff4cf
d78d9dd
7ac0cfd
734e113
ab9fffd
71d6536
9ed511b
c1c9b98
f667267
f9a6a96
015d64f
b6188c1
b9a3548
ae31d82
195dc1d
58060cc
b99d973
bacfc1e
afa89d4
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
Started at: 04:40:02
norbert, 0.001, 256
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 0.490340878367424 - MAE: 0.5502402056555946
Validation loss : 0.2153187460369534 - MAE: 0.34847071386417267
Epoch: 1
Training loss: 0.20058341324329376 - MAE: 0.3423091295468801
Validation loss : 0.17609875235292646 - MAE: 0.3207490493525149
Epoch: 2
Training loss: 0.17480310261249543 - MAE: 0.31785240619760485
Validation loss : 0.17104972898960114 - MAE: 0.31704982503328616
Epoch: 3
Training loss: 0.16844426095485687 - MAE: 0.31094802683732065
Validation loss : 0.16871228482988146 - MAE: 0.3151468612097043
Epoch: 4
Training loss: 0.16473501205444335 - MAE: 0.3070631323882683
Validation loss : 0.16664770245552063 - MAE: 0.3130509722987033
Epoch: 5
Training loss: 0.16249913841485977 - MAE: 0.30390396654148055
Validation loss : 0.16460454960664114 - MAE: 0.3108272246604274
Epoch: 6
Training loss: 0.16045642793178558 - MAE: 0.3025153689886295
Validation loss : 0.16342525018586052 - MAE: 0.30940071952857345
Epoch: 7
Training loss: 0.15999574840068817 - MAE: 0.3011067249123912
Validation loss : 0.16275884045494926 - MAE: 0.3085626349629923
Epoch: 8
Training loss: 0.15708923935890198 - MAE: 0.29919283959579873
Validation loss : 0.1618599063820309 - MAE: 0.3077045065004142
Epoch: 9
Training loss: 0.1566512778401375 - MAE: 0.29883795388199974
Validation loss : 0.16123364369074503 - MAE: 0.3070243111305445
Epoch: 10
Training loss: 0.1565144717693329 - MAE: 0.2992467122905806
Validation loss : 0.1607403705517451 - MAE: 0.3066506261102796
Epoch: 11
Training loss: 0.15441132932901383 - MAE: 0.2968694001405616
Validation loss : 0.1615937219725715 - MAE: 0.30805963680583925
Epoch: 12
Training loss: 0.15473220765590667 - MAE: 0.2973775166487022
Validation loss : 0.1601952744854821 - MAE: 0.30593868950315056
Epoch: 13
Training loss: 0.15416186094284057 - MAE: 0.2955810503214238
Validation loss : 0.1606642852226893 - MAE: 0.30693562115452544
Epoch: 14
Training loss: 0.1523517221212387 - MAE: 0.29359860648947156
Validation loss : 0.15979144473870596 - MAE: 0.3059113352899167
Epoch: 15
Training loss: 0.15219720542430878 - MAE: 0.29356666946971216
Validation loss : 0.15870776772499084 - MAE: 0.30447905617139626
Epoch: 16
Training loss: 0.15407244622707367 - MAE: 0.2968844009018846
Validation loss : 0.15995393693447113 - MAE: 0.305935636680547
Epoch: 17
Training loss: 0.15287074476480483 - MAE: 0.2946156906987487
Validation loss : 0.1583722018533283 - MAE: 0.30404090762595287
Epoch: 18
Training loss: 0.15273998081684112 - MAE: 0.2950003350810963
Validation loss : 0.15665675865279305 - MAE: 0.3017202135107541