File size: 4,206 Bytes
415e1d6
0625bb2
7d08e24
557aa89
450b003
94d6b3c
4f21eea
6a81b5a
5ae1f34
0b261b6
64c82db
bf2ade6
5218d77
1f36c26
9a66345
7b89b3f
f6ba68d
2a24f47
9d1c1f7
7261512
d01a3d5
d560ed1
b67e2d7
25ca10f
ed1681d
1afcba2
88b3981
5433b50
922374e
5a1c0c6
3b92628
ba2995f
3099c28
d2a9fed
91709a5
3c755dc
9f5271e
5e9a2ba
070cc17
a1d6b8d
f6d2bf1
acfd692
e0c788c
0729b84
3af0a40
012213c
eb44138
8a6167a
1ea0a3b
4ccabd1
a11eff6
9c17bda
1a1685c
76639f7
aa27110
6f639bc
cb06b25
a234af1
9bda0b1
9588a83
7e39a94
5a5a2ab
1bdac7b
1bc6c22
c5eae9e
105067f
2e7714c
332948e
e7e7a79
c983ea2
628f7e7
364616a
85c5d50
20d007f
c7a03ea
3ad0b5b
e342779
31d504b
7968def
2c9f269
9a8fd67
0af32e1
0ea38b3
a07c6bd
59047c5
8c4a632
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
Started at: 14:22:52
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 0.6967193094583658 - MAE: 0.6750990625330823
Validation loss : 0.5263436496257782 - MAE: 0.613448569191522
Epoch: 1
Training loss: 0.26826212727106535 - MAE: 0.4087051125792896
Validation loss : 0.22296061515808105 - MAE: 0.3555378972565709
Epoch: 2
Training loss: 0.2124968572304799 - MAE: 0.3476390507866737
Validation loss : 0.20461321175098418 - MAE: 0.3532590132614058
Epoch: 3
Training loss: 0.18429063145930952 - MAE: 0.32677963264803717
Validation loss : 0.17646469473838805 - MAE: 0.3203760276514041
Epoch: 4
Training loss: 0.17768853444319505 - MAE: 0.31684026520272845
Validation loss : 0.17874470055103303 - MAE: 0.3257858562177419
Epoch: 5
Training loss: 0.17304767782871538 - MAE: 0.3139755194179104
Validation loss : 0.1718216747045517 - MAE: 0.31812428100953294
Epoch: 6
Training loss: 0.17041566165593955 - MAE: 0.31176170900212385
Validation loss : 0.17110097408294678 - MAE: 0.3175205683975872
Epoch: 7
Training loss: 0.16822290764405176 - MAE: 0.3093251221653364
Validation loss : 0.17010048031806946 - MAE: 0.3165555384963895
Epoch: 8
Training loss: 0.16749977606993455 - MAE: 0.309235729717861
Validation loss : 0.1681269407272339 - MAE: 0.31405053463780375
Epoch: 9
Training loss: 0.1660219534085347 - MAE: 0.30707150650440435
Validation loss : 0.16768858432769776 - MAE: 0.3134797613244496
Epoch: 10
Training loss: 0.16200380256542793 - MAE: 0.3036754525218348
Validation loss : 0.1665838599205017 - MAE: 0.3121892924671111
Epoch: 11
Training loss: 0.16332647777520692 - MAE: 0.3039458888319623
Validation loss : 0.16546964943408965 - MAE: 0.3107545377167428
Epoch: 12
Training loss: 0.16287456223597893 - MAE: 0.30448042997749564
Validation loss : 0.16524331867694855 - MAE: 0.3106786227351389
Epoch: 13
Training loss: 0.15990304029904878 - MAE: 0.30016709856440854
Validation loss : 0.16489878594875335 - MAE: 0.31031377115036685
Epoch: 14
Training loss: 0.15767766993779403 - MAE: 0.2997244083849813
Validation loss : 0.16339748501777648 - MAE: 0.3083185140265562
Epoch: 15
Training loss: 0.15951268031046942 - MAE: 0.29912324704552734
Validation loss : 0.16375492811203002 - MAE: 0.3090144460164269
Epoch: 16
Training loss: 0.15623677808504838 - MAE: 0.2973088492735953
Validation loss : 0.16299248337745667 - MAE: 0.30820454237494843
Epoch: 17
Training loss: 0.1564337038076841 - MAE: 0.29861382568944933
Validation loss : 0.16230234503746033 - MAE: 0.30725784051523525
Epoch: 18
Training loss: 0.15549989617787874 - MAE: 0.2960598878509324
Validation loss : 0.16213322579860687 - MAE: 0.3070436707425703
Epoch: 19
Training loss: 0.1569853677199437 - MAE: 0.29836095928185685
Validation loss : 0.16146726608276368 - MAE: 0.30629696980134635
Epoch: 20
Training loss: 0.1547408069555576 - MAE: 0.2970990607700864
Validation loss : 0.16105090379714965 - MAE: 0.3058310992073678
Epoch: 21
Training loss: 0.154588805941435 - MAE: 0.2957786660002471
Validation loss : 0.16075867414474487 - MAE: 0.30553690240829345
Epoch: 22
Training loss: 0.153683405656081 - MAE: 0.29670572424115915
Validation loss : 0.16087317764759063 - MAE: 0.3056201053962598
Epoch: 23
Training loss: 0.15514581363934737 - MAE: 0.2955321868620246
Validation loss : 0.16066641211509705 - MAE: 0.30531659865767746
Epoch: 24
Training loss: 0.15376647619100717 - MAE: 0.2947861379473795
Validation loss : 0.1603192001581192 - MAE: 0.3047779291624614
Epoch: 25
Training loss: 0.15424712231526008 - MAE: 0.295517101740558
Validation loss : 0.16027258038520814 - MAE: 0.30490525127423435
Epoch: 26
Training loss: 0.1534990152487388 - MAE: 0.2947872614603403
Validation loss : 0.15944191813468933 - MAE: 0.3040211334248351
Epoch: 27
Training loss: 0.15433337367497957 - MAE: 0.29452155782914224
Validation loss : 0.1589920550584793 - MAE: 0.3033883763699271