File size: 4,210 Bytes
2d4dd88
172a071
acfbf06
747f9b0
7966e32
19b9445
a9a19a6
1bd8f8a
0a179ba
12cd077
bd89888
429c852
6e4b5e3
8cb5b06
b2aee6e
ca6924e
096bd8e
4c0d99a
8337c0e
c7ce869
0914010
ee230b9
8b59da9
e3e1557
98a5ea7
c6ba2fc
4b68af1
c30e6d9
cb93f28
9f998e7
509a543
ae5ef95
c9e236b
c5a2f71
d81e4f0
e4f9fb3
d1378a2
4ad696e
5d2e438
0fcec26
9f6f7f5
bcbe884
bd63c21
3d12e41
909eb47
2a4df92
f0a2de1
54d89f0
1e6aad7
295dfe9
39c6e20
ebdc6b7
7825e7d
0dd94ea
0233eb5
897faff
e9ffbdc
1079f97
193f1b0
2800774
2994467
a62c4de
c5e015a
9cfb817
2ed6e0e
3acefdc
f63941e
aeb1b1c
2980eb6
4813490
53d8184
7d75378
cfeaf59
9e8a062
4b44244
b3a13c5
a97196f
2f229ec
b5c14f1
f495bfd
f42bc4e
3afeadd
8ab3362
6fff3c8
62f8d9d
1abb531
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
Started at: 12:33:08
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
Epoch: 0
Training loss: 0.34034032225608823 - MAE: 0.45751100328425043
Validation loss : 0.20192567838562858 - MAE: 0.3443028131702277
Epoch: 1
Training loss: 0.21138036668300628 - MAE: 0.35259469495701584
Validation loss : 0.19307727449470097 - MAE: 0.3411017983370112
Epoch: 2
Training loss: 0.1951470822095871 - MAE: 0.3371726994373718
Validation loss : 0.1826503872871399 - MAE: 0.3290534102707088
Epoch: 3
Training loss: 0.1880119514465332 - MAE: 0.330053510216342
Validation loss : 0.17956634031401741 - MAE: 0.32711545778572254
Epoch: 4
Training loss: 0.1804283791780472 - MAE: 0.32134730454152893
Validation loss : 0.176905600561036 - MAE: 0.32432533000410785
Epoch: 5
Training loss: 0.18152455627918243 - MAE: 0.3227039616721611
Validation loss : 0.17382549080583784 - MAE: 0.3212278624694203
Epoch: 6
Training loss: 0.17641499400138855 - MAE: 0.3178502426835815
Validation loss : 0.17528054945998722 - MAE: 0.3241598988895522
Epoch: 7
Training loss: 0.1722068202495575 - MAE: 0.31442449621456453
Validation loss : 0.1724375370475981 - MAE: 0.32140079938239263
Epoch: 8
Training loss: 0.16665148079395295 - MAE: 0.30963764222638895
Validation loss : 0.1672506266170078 - MAE: 0.31489036384905245
Epoch: 9
Training loss: 0.16707129299640655 - MAE: 0.3100291273814063
Validation loss : 0.16691898637347752 - MAE: 0.31496471450269947
Epoch: 10
Training loss: 0.1675616580247879 - MAE: 0.30764175755177864
Validation loss : 0.1650147471163008 - MAE: 0.31304393625062915
Epoch: 11
Training loss: 0.16691500306129456 - MAE: 0.3078508740458523
Validation loss : 0.16393668784035575 - MAE: 0.31154094680405375
Epoch: 12
Training loss: 0.1637336140871048 - MAE: 0.30385438993037184
Validation loss : 0.16229701538880667 - MAE: 0.30975237715702897
Epoch: 13
Training loss: 0.1618727833032608 - MAE: 0.3037893453784594
Validation loss : 0.16030184262328678 - MAE: 0.30714003484399144
Epoch: 14
Training loss: 0.16216390013694762 - MAE: 0.30370377205429017
Validation loss : 0.1587020523018307 - MAE: 0.30477575013081115
Epoch: 15
Training loss: 0.16157052159309387 - MAE: 0.30328576923920736
Validation loss : 0.15763747692108154 - MAE: 0.30343286374299505
Epoch: 16
Training loss: 0.1586935031414032 - MAE: 0.2999908923261831
Validation loss : 0.1570943941672643 - MAE: 0.30202990273409386
Epoch: 17
Training loss: 0.15946080952882766 - MAE: 0.30086799079545556
Validation loss : 0.15687472952736747 - MAE: 0.3010396265266053
Epoch: 18
Training loss: 0.15581952095031737 - MAE: 0.2973425589462948
Validation loss : 0.15603283047676086 - MAE: 0.30100286245604474
Epoch: 19
Training loss: 0.15706016898155212 - MAE: 0.2994781823556224
Validation loss : 0.15608555575211844 - MAE: 0.3011623471741622
Epoch: 20
Training loss: 0.15702145874500276 - MAE: 0.2992172543463727
Validation loss : 0.15507571233643425 - MAE: 0.2998725774790534
Epoch: 21
Training loss: 0.15234391868114472 - MAE: 0.29315056716991045
Validation loss : 0.15489690999190012 - MAE: 0.29869079940078813
Epoch: 22
Training loss: 0.153500095307827 - MAE: 0.29595609632775555
Validation loss : 0.15607044597466788 - MAE: 0.2988202550812423
Epoch: 23
Training loss: 0.1513380992412567 - MAE: 0.2943754520786924
Validation loss : 0.15447033445040384 - MAE: 0.29746065098913443
Epoch: 24
Training loss: 0.15459679573774338 - MAE: 0.29592691579298513
Validation loss : 0.1555763234694799 - MAE: 0.298155070237502
Epoch: 25
Training loss: 0.15357773929834365 - MAE: 0.29463498551553785
Validation loss : 0.15434908039040035 - MAE: 0.296779802279584
Epoch: 26
Training loss: 0.15193950653076171 - MAE: 0.2929094511418085
Validation loss : 0.15472973386446634 - MAE: 0.297041940424016
Epoch: 27
Training loss: 0.14988958299160005 - MAE: 0.2907063248230569
Validation loss : 0.15451522833771175 - MAE: 0.296753490150844