File size: 3,139 Bytes
797a72b
8b51bb6
882fa36
c15ab4c
1bd332f
742b96a
4465920
6024120
97bf903
8d0db44
6d2e601
1ac98e0
461aac8
e939fb4
8340d3a
f0ed3e9
9c88970
2fc70a1
21f0e15
7cd3976
2b34a6b
317869f
1602c7d
f8d8c9e
7494fb7
30f96cc
81d92ac
2813fa7
666b9a3
7bbd829
3630a81
e012ec3
a0243b7
4a08212
58e9801
8efa922
5d13d84
f966bf2
d5f9f9d
ca60a80
890ba4f
ffb81cd
bc32ea2
2c85c82
07a7942
78fa631
2906806
63d877a
3e75a18
cbe04e0
ca5aee7
bd2add5
5ece256
0d2e2b8
72c80fd
0682c6d
f1463fa
2c3b967
13c59eb
db2f4fb
b7f62d6
b564e7e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
Started at: 10:46:39
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 0.4224580281972885 - MAE: 0.49800078344951854
Validation loss : 0.18566939760656917 - MAE: 0.32860604518901193
Epoch: 1
Training loss: 0.18062527552247049 - MAE: 0.3249083793427436
Validation loss : 0.17773295325391433 - MAE: 0.32505375557728605
Epoch: 2
Training loss: 0.17239214196801186 - MAE: 0.3161783884766083
Validation loss : 0.1721238993546542 - MAE: 0.31926727109425135
Epoch: 3
Training loss: 0.16808241829276085 - MAE: 0.3116257252131245
Validation loss : 0.16848680245525696 - MAE: 0.31521207737718204
Epoch: 4
Training loss: 0.16556511253118514 - MAE: 0.30872840319635675
Validation loss : 0.16570698601358078 - MAE: 0.31138876910079605
Epoch: 5
Training loss: 0.16236369609832763 - MAE: 0.3063052400904153
Validation loss : 0.16389068610527935 - MAE: 0.30842703017799394
Epoch: 6
Training loss: 0.15821502283215522 - MAE: 0.3002932636657733
Validation loss : 0.16259106204790227 - MAE: 0.3064878174082187
Epoch: 7
Training loss: 0.15610535070300102 - MAE: 0.29970597196416643
Validation loss : 0.16118212920777938 - MAE: 0.30487487035579575
Epoch: 8
Training loss: 0.15775163292884828 - MAE: 0.3012530324315296
Validation loss : 0.16053338874788844 - MAE: 0.30467935002674695
Epoch: 9
Training loss: 0.15300563350319862 - MAE: 0.2966342830465275
Validation loss : 0.1598155660664334 - MAE: 0.3024349783524495
Epoch: 10
Training loss: 0.15308804035186768 - MAE: 0.29689653867839455
Validation loss : 0.15887342218090506 - MAE: 0.30158871157294653
Epoch: 11
Training loss: 0.15205074325203896 - MAE: 0.2968189546038619
Validation loss : 0.15820855384363847 - MAE: 0.3009622242442208
Epoch: 12
Training loss: 0.14986242949962617 - MAE: 0.29354556482882427
Validation loss : 0.15779440937673345 - MAE: 0.3011403656557578
Epoch: 13
Training loss: 0.14915254317224025 - MAE: 0.29414236445220476
Validation loss : 0.1570692671572461 - MAE: 0.29919229552027593
Epoch: 14
Training loss: 0.1507771185040474 - MAE: 0.2942755191435603
Validation loss : 0.15648663569899166 - MAE: 0.2990671428995643
Epoch: 15
Training loss: 0.14960825718939305 - MAE: 0.29282880951675244
Validation loss : 0.15587416028275208 - MAE: 0.29818727256321803
Epoch: 16
Training loss: 0.14761282175779342 - MAE: 0.290322619453174
Validation loss : 0.15559193623416565 - MAE: 0.2986693999701733
Epoch: 17
Training loss: 0.14804843626916409 - MAE: 0.2924950026363286
Validation loss : 0.15508160126559875 - MAE: 0.2970730023460469
Epoch: 18
Training loss: 0.14751567400991916 - MAE: 0.28970414097885494
Validation loss : 0.15513849389903686 - MAE: 0.2973691548251751
Epoch: 19
Training loss: 0.1472933314740658 - MAE: 0.2901643469721373
Validation loss : 0.15472981070770936 - MAE: 0.29793042253066404