File size: 4,581 Bytes
8ecdf1a
85b6090
e3c6bc8
84b320c
f0d1e38
25a1eeb
74bffdd
cd3af81
55d7b2d
9c9c762
f656e6b
0a5d8f0
0ebdc8c
cb178ff
711c898
f5aa86d
9e46af4
00d2388
d75a248
4cf1834
c4920e5
15a02c7
6461657
73d6e8e
700ac7c
1da0fe1
61b2725
0d6baea
cc82b0c
0ff25b2
458124b
2a0d03d
390e53e
f1e3eb7
bb1acca
f366bc5
381e209
5ac883f
41ffb32
54eb5da
1b31ab2
0c83ab5
60c4a64
ad072a4
aafd56e
6ec2e7d
6d1e465
6918809
dcbb528
b82ce9c
c3635dd
b75a562
7a4a3d5
5c16c4b
65080ab
34dbfc1
a458191
59915cc
8377a8e
1a87211
866acd0
6788872
7c84aff
6f152f8
557e3f4
c3eb03f
5fea102
fe115eb
2f261be
89b20bc
ad41e2d
6944f69
08a7bb0
2970e05
9a70e81
4540483
da65d5a
1084f61
88ac2bd
3d7a869
ab2be0d
d824807
210abf0
fd70ab0
9364dc5
ab5c667
4c091a9
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
Started at: 00:24:23
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 0.6361208361387253 - MAE: 0.6347873231265873
Validation loss : 0.33865517377853394 - MAE: 0.4740741020180635
Epoch: 1
Training loss: 0.21345499098300935 - MAE: 0.3540912858998663
Validation loss : 0.18473448190424177 - MAE: 0.3282540270842396
Epoch: 2
Training loss: 0.18606460213661194 - MAE: 0.32889828686188877
Validation loss : 0.18045285178555381 - MAE: 0.32713884374037794
Epoch: 3
Training loss: 0.17846946954727172 - MAE: 0.31904566117043975
Validation loss : 0.17773988511827257 - MAE: 0.32482577310153565
Epoch: 4
Training loss: 0.17139309346675874 - MAE: 0.31441827165194275
Validation loss : 0.17418616182274288 - MAE: 0.32107063596093277
Epoch: 5
Training loss: 0.17030848145484925 - MAE: 0.31198532199217943
Validation loss : 0.17218486799134147 - MAE: 0.3190577970415032
Epoch: 6
Training loss: 0.16878579378128053 - MAE: 0.31078672095469284
Validation loss : 0.17055005000697243 - MAE: 0.3174862826881529
Epoch: 7
Training loss: 0.1679440525174141 - MAE: 0.31018192416787727
Validation loss : 0.16925331950187683 - MAE: 0.31617688409581507
Epoch: 8
Training loss: 0.16459781765937806 - MAE: 0.3076978685246546
Validation loss : 0.16737455626328787 - MAE: 0.313686903818875
Epoch: 9
Training loss: 0.1616947764158249 - MAE: 0.3042913169405816
Validation loss : 0.16636485523647732 - MAE: 0.3125500280647287
Started at: 01:47:29
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 0.6361208361387253 - MAE: 0.6347873231265873
Validation loss : 0.33865517377853394 - MAE: 0.4740741020180635
Epoch: 1
Training loss: 0.21345499098300935 - MAE: 0.3540912858998663
Validation loss : 0.18473448190424177 - MAE: 0.3282540270842396
Epoch: 2
Training loss: 0.18606460213661194 - MAE: 0.32889828686188877
Validation loss : 0.18045285178555381 - MAE: 0.32713884374037794
Epoch: 3
Training loss: 0.17846946954727172 - MAE: 0.31904566117043975
Validation loss : 0.17773988511827257 - MAE: 0.32482577310153565
Epoch: 4
Training loss: 0.17139309346675874 - MAE: 0.31441827165194275
Validation loss : 0.17418616182274288 - MAE: 0.32107063596093277
Epoch: 5
Training loss: 0.17030848145484925 - MAE: 0.31198532199217943
Validation loss : 0.17218486799134147 - MAE: 0.3190577970415032
Epoch: 6
Training loss: 0.16878579378128053 - MAE: 0.31078672095469284
Validation loss : 0.17055005000697243 - MAE: 0.3174862826881529
Epoch: 7
Training loss: 0.1679440525174141 - MAE: 0.31018192416787727
Validation loss : 0.16925331950187683 - MAE: 0.31617688409581507
Epoch: 8
Training loss: 0.16459781765937806 - MAE: 0.3076978685246546
Validation loss : 0.16737455626328787 - MAE: 0.313686903818875
Epoch: 9
Training loss: 0.1616947764158249 - MAE: 0.3042913169405816
Validation loss : 0.16636485523647732 - MAE: 0.3125500280647287
Epoch: 10
Training loss: 0.1614231312274933 - MAE: 0.30281138726351875
Validation loss : 0.16527396109369066 - MAE: 0.310935884275779
Epoch: 11
Training loss: 0.15981832027435303 - MAE: 0.3036158805084463
Validation loss : 0.16393334501319462 - MAE: 0.3090806138434242
Epoch: 12
Training loss: 0.1593213653564453 - MAE: 0.30149370061446074
Validation loss : 0.1633289953072866 - MAE: 0.3085658493030115
Epoch: 13
Training loss: 0.15871796071529387 - MAE: 0.30060354026448305
Validation loss : 0.16265849934683907 - MAE: 0.30747698672849366
Epoch: 14
Training loss: 0.15790118873119355 - MAE: 0.29930967856529156
Validation loss : 0.1618209034204483 - MAE: 0.3064716457801814
Epoch: 15
Training loss: 0.15594214886426927 - MAE: 0.2976610109442401
Validation loss : 0.1610150718026691 - MAE: 0.30535239027482275
Epoch: 16
Training loss: 0.15538664102554323 - MAE: 0.29795636864916086
Validation loss : 0.16068673796123928 - MAE: 0.3052027362150315
Epoch: 17
Training loss: 0.15572097659111023 - MAE: 0.2970189372002837