dfanswerrocket commited on
Commit
72aea89
1 Parent(s): 770b27f

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +101 -101
README.md CHANGED
@@ -14,7 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
- - Loss: 1.9029
18
 
19
  ## Model description
20
 
@@ -45,106 +45,106 @@ The following hyperparameters were used during training:
45
 
46
  | Training Loss | Epoch | Step | Validation Loss |
47
  |:-------------:|:-----:|:----:|:---------------:|
48
- | No log | 1.0 | 71 | 8.8951 |
49
- | No log | 2.0 | 142 | 8.8532 |
50
- | No log | 3.0 | 213 | 8.7305 |
51
- | No log | 4.0 | 284 | 8.5186 |
52
- | No log | 5.0 | 355 | 8.3212 |
53
- | No log | 6.0 | 426 | 8.1309 |
54
- | No log | 7.0 | 497 | 7.9420 |
55
- | 8.6328 | 8.0 | 568 | 7.7689 |
56
- | 8.6328 | 9.0 | 639 | 7.6050 |
57
- | 8.6328 | 10.0 | 710 | 7.4483 |
58
- | 8.6328 | 11.0 | 781 | 7.2902 |
59
- | 8.6328 | 12.0 | 852 | 7.1421 |
60
- | 8.6328 | 13.0 | 923 | 6.9918 |
61
- | 8.6328 | 14.0 | 994 | 6.8501 |
62
- | 7.5715 | 15.0 | 1065 | 6.7115 |
63
- | 7.5715 | 16.0 | 1136 | 6.5774 |
64
- | 7.5715 | 17.0 | 1207 | 6.4431 |
65
- | 7.5715 | 18.0 | 1278 | 6.3216 |
66
- | 7.5715 | 19.0 | 1349 | 6.1973 |
67
- | 7.5715 | 20.0 | 1420 | 6.0735 |
68
- | 7.5715 | 21.0 | 1491 | 5.9483 |
69
- | 6.6466 | 22.0 | 1562 | 5.8385 |
70
- | 6.6466 | 23.0 | 1633 | 5.7199 |
71
- | 6.6466 | 24.0 | 1704 | 5.6047 |
72
- | 6.6466 | 25.0 | 1775 | 5.5008 |
73
- | 6.6466 | 26.0 | 1846 | 5.3816 |
74
- | 6.6466 | 27.0 | 1917 | 5.2802 |
75
- | 6.6466 | 28.0 | 1988 | 5.1727 |
76
- | 5.8483 | 29.0 | 2059 | 5.0741 |
77
- | 5.8483 | 30.0 | 2130 | 4.9724 |
78
- | 5.8483 | 31.0 | 2201 | 4.8772 |
79
- | 5.8483 | 32.0 | 2272 | 4.7812 |
80
- | 5.8483 | 33.0 | 2343 | 4.6847 |
81
- | 5.8483 | 34.0 | 2414 | 4.5938 |
82
- | 5.8483 | 35.0 | 2485 | 4.5092 |
83
- | 5.1726 | 36.0 | 2556 | 4.4214 |
84
- | 5.1726 | 37.0 | 2627 | 4.3330 |
85
- | 5.1726 | 38.0 | 2698 | 4.2479 |
86
- | 5.1726 | 39.0 | 2769 | 4.1682 |
87
- | 5.1726 | 40.0 | 2840 | 4.0822 |
88
- | 5.1726 | 41.0 | 2911 | 4.0075 |
89
- | 5.1726 | 42.0 | 2982 | 3.9296 |
90
- | 4.5503 | 43.0 | 3053 | 3.8511 |
91
- | 4.5503 | 44.0 | 3124 | 3.7752 |
92
- | 4.5503 | 45.0 | 3195 | 3.7009 |
93
- | 4.5503 | 46.0 | 3266 | 3.6335 |
94
- | 4.5503 | 47.0 | 3337 | 3.5647 |
95
- | 4.5503 | 48.0 | 3408 | 3.4937 |
96
- | 4.5503 | 49.0 | 3479 | 3.4242 |
97
- | 4.009 | 50.0 | 3550 | 3.3637 |
98
- | 4.009 | 51.0 | 3621 | 3.3031 |
99
- | 4.009 | 52.0 | 3692 | 3.2395 |
100
- | 4.009 | 53.0 | 3763 | 3.1766 |
101
- | 4.009 | 54.0 | 3834 | 3.1246 |
102
- | 4.009 | 55.0 | 3905 | 3.0673 |
103
- | 4.009 | 56.0 | 3976 | 3.0090 |
104
- | 3.5477 | 57.0 | 4047 | 2.9559 |
105
- | 3.5477 | 58.0 | 4118 | 2.9098 |
106
- | 3.5477 | 59.0 | 4189 | 2.8603 |
107
- | 3.5477 | 60.0 | 4260 | 2.8132 |
108
- | 3.5477 | 61.0 | 4331 | 2.7637 |
109
- | 3.5477 | 62.0 | 4402 | 2.7219 |
110
- | 3.5477 | 63.0 | 4473 | 2.6734 |
111
- | 3.1822 | 64.0 | 4544 | 2.6286 |
112
- | 3.1822 | 65.0 | 4615 | 2.5867 |
113
- | 3.1822 | 66.0 | 4686 | 2.5491 |
114
- | 3.1822 | 67.0 | 4757 | 2.5076 |
115
- | 3.1822 | 68.0 | 4828 | 2.4720 |
116
- | 3.1822 | 69.0 | 4899 | 2.4349 |
117
- | 3.1822 | 70.0 | 4970 | 2.4026 |
118
- | 2.8647 | 71.0 | 5041 | 2.3661 |
119
- | 2.8647 | 72.0 | 5112 | 2.3343 |
120
- | 2.8647 | 73.0 | 5183 | 2.3081 |
121
- | 2.8647 | 74.0 | 5254 | 2.2751 |
122
- | 2.8647 | 75.0 | 5325 | 2.2461 |
123
- | 2.8647 | 76.0 | 5396 | 2.2165 |
124
- | 2.8647 | 77.0 | 5467 | 2.1938 |
125
- | 2.6306 | 78.0 | 5538 | 2.1681 |
126
- | 2.6306 | 79.0 | 5609 | 2.1438 |
127
- | 2.6306 | 80.0 | 5680 | 2.1232 |
128
- | 2.6306 | 81.0 | 5751 | 2.1022 |
129
- | 2.6306 | 82.0 | 5822 | 2.0800 |
130
- | 2.6306 | 83.0 | 5893 | 2.0605 |
131
- | 2.6306 | 84.0 | 5964 | 2.0427 |
132
- | 2.4436 | 85.0 | 6035 | 2.0273 |
133
- | 2.4436 | 86.0 | 6106 | 2.0109 |
134
- | 2.4436 | 87.0 | 6177 | 1.9975 |
135
- | 2.4436 | 88.0 | 6248 | 1.9827 |
136
- | 2.4436 | 89.0 | 6319 | 1.9700 |
137
- | 2.4436 | 90.0 | 6390 | 1.9584 |
138
- | 2.4436 | 91.0 | 6461 | 1.9502 |
139
- | 2.3168 | 92.0 | 6532 | 1.9394 |
140
- | 2.3168 | 93.0 | 6603 | 1.9307 |
141
- | 2.3168 | 94.0 | 6674 | 1.9221 |
142
- | 2.3168 | 95.0 | 6745 | 1.9167 |
143
- | 2.3168 | 96.0 | 6816 | 1.9122 |
144
- | 2.3168 | 97.0 | 6887 | 1.9078 |
145
- | 2.3168 | 98.0 | 6958 | 1.9055 |
146
- | 2.2563 | 99.0 | 7029 | 1.9039 |
147
- | 2.2563 | 100.0 | 7100 | 1.9029 |
148
 
149
 
150
  ### Framework versions
 
14
 
15
  This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the None dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Loss: 1.9992
18
 
19
  ## Model description
20
 
 
45
 
46
  | Training Loss | Epoch | Step | Validation Loss |
47
  |:-------------:|:-----:|:----:|:---------------:|
48
+ | No log | 1.0 | 67 | 8.8964 |
49
+ | No log | 2.0 | 134 | 8.8550 |
50
+ | No log | 3.0 | 201 | 8.7423 |
51
+ | No log | 4.0 | 268 | 8.5343 |
52
+ | No log | 5.0 | 335 | 8.3243 |
53
+ | No log | 6.0 | 402 | 8.1392 |
54
+ | No log | 7.0 | 469 | 7.9616 |
55
+ | 8.6114 | 8.0 | 536 | 7.7924 |
56
+ | 8.6114 | 9.0 | 603 | 7.6305 |
57
+ | 8.6114 | 10.0 | 670 | 7.4707 |
58
+ | 8.6114 | 11.0 | 737 | 7.3065 |
59
+ | 8.6114 | 12.0 | 804 | 7.1600 |
60
+ | 8.6114 | 13.0 | 871 | 7.0228 |
61
+ | 8.6114 | 14.0 | 938 | 6.8804 |
62
+ | 7.5409 | 15.0 | 1005 | 6.7334 |
63
+ | 7.5409 | 16.0 | 1072 | 6.6021 |
64
+ | 7.5409 | 17.0 | 1139 | 6.4789 |
65
+ | 7.5409 | 18.0 | 1206 | 6.3473 |
66
+ | 7.5409 | 19.0 | 1273 | 6.2252 |
67
+ | 7.5409 | 20.0 | 1340 | 6.1058 |
68
+ | 7.5409 | 21.0 | 1407 | 5.9892 |
69
+ | 7.5409 | 22.0 | 1474 | 5.8674 |
70
+ | 6.6051 | 23.0 | 1541 | 5.7496 |
71
+ | 6.6051 | 24.0 | 1608 | 5.6393 |
72
+ | 6.6051 | 25.0 | 1675 | 5.5244 |
73
+ | 6.6051 | 26.0 | 1742 | 5.4279 |
74
+ | 6.6051 | 27.0 | 1809 | 5.3221 |
75
+ | 6.6051 | 28.0 | 1876 | 5.2126 |
76
+ | 6.6051 | 29.0 | 1943 | 5.1221 |
77
+ | 5.8003 | 30.0 | 2010 | 5.0178 |
78
+ | 5.8003 | 31.0 | 2077 | 4.9183 |
79
+ | 5.8003 | 32.0 | 2144 | 4.8303 |
80
+ | 5.8003 | 33.0 | 2211 | 4.7328 |
81
+ | 5.8003 | 34.0 | 2278 | 4.6467 |
82
+ | 5.8003 | 35.0 | 2345 | 4.5548 |
83
+ | 5.8003 | 36.0 | 2412 | 4.4697 |
84
+ | 5.8003 | 37.0 | 2479 | 4.3860 |
85
+ | 5.0905 | 38.0 | 2546 | 4.2980 |
86
+ | 5.0905 | 39.0 | 2613 | 4.2172 |
87
+ | 5.0905 | 40.0 | 2680 | 4.1351 |
88
+ | 5.0905 | 41.0 | 2747 | 4.0635 |
89
+ | 5.0905 | 42.0 | 2814 | 3.9834 |
90
+ | 5.0905 | 43.0 | 2881 | 3.9091 |
91
+ | 5.0905 | 44.0 | 2948 | 3.8376 |
92
+ | 4.481 | 45.0 | 3015 | 3.7662 |
93
+ | 4.481 | 46.0 | 3082 | 3.7011 |
94
+ | 4.481 | 47.0 | 3149 | 3.6335 |
95
+ | 4.481 | 48.0 | 3216 | 3.5671 |
96
+ | 4.481 | 49.0 | 3283 | 3.5011 |
97
+ | 4.481 | 50.0 | 3350 | 3.4388 |
98
+ | 4.481 | 51.0 | 3417 | 3.3746 |
99
+ | 4.481 | 52.0 | 3484 | 3.3151 |
100
+ | 3.9521 | 53.0 | 3551 | 3.2551 |
101
+ | 3.9521 | 54.0 | 3618 | 3.1943 |
102
+ | 3.9521 | 55.0 | 3685 | 3.1410 |
103
+ | 3.9521 | 56.0 | 3752 | 3.0885 |
104
+ | 3.9521 | 57.0 | 3819 | 3.0384 |
105
+ | 3.9521 | 58.0 | 3886 | 2.9890 |
106
+ | 3.9521 | 59.0 | 3953 | 2.9376 |
107
+ | 3.5177 | 60.0 | 4020 | 2.8906 |
108
+ | 3.5177 | 61.0 | 4087 | 2.8406 |
109
+ | 3.5177 | 62.0 | 4154 | 2.7951 |
110
+ | 3.5177 | 63.0 | 4221 | 2.7590 |
111
+ | 3.5177 | 64.0 | 4288 | 2.7136 |
112
+ | 3.5177 | 65.0 | 4355 | 2.6725 |
113
+ | 3.5177 | 66.0 | 4422 | 2.6343 |
114
+ | 3.5177 | 67.0 | 4489 | 2.5941 |
115
+ | 3.1601 | 68.0 | 4556 | 2.5563 |
116
+ | 3.1601 | 69.0 | 4623 | 2.5241 |
117
+ | 3.1601 | 70.0 | 4690 | 2.4894 |
118
+ | 3.1601 | 71.0 | 4757 | 2.4552 |
119
+ | 3.1601 | 72.0 | 4824 | 2.4227 |
120
+ | 3.1601 | 73.0 | 4891 | 2.3942 |
121
+ | 3.1601 | 74.0 | 4958 | 2.3678 |
122
+ | 2.8815 | 75.0 | 5025 | 2.3362 |
123
+ | 2.8815 | 76.0 | 5092 | 2.3100 |
124
+ | 2.8815 | 77.0 | 5159 | 2.2851 |
125
+ | 2.8815 | 78.0 | 5226 | 2.2570 |
126
+ | 2.8815 | 79.0 | 5293 | 2.2346 |
127
+ | 2.8815 | 80.0 | 5360 | 2.2155 |
128
+ | 2.8815 | 81.0 | 5427 | 2.1933 |
129
+ | 2.8815 | 82.0 | 5494 | 2.1714 |
130
+ | 2.6556 | 83.0 | 5561 | 2.1551 |
131
+ | 2.6556 | 84.0 | 5628 | 2.1381 |
132
+ | 2.6556 | 85.0 | 5695 | 2.1203 |
133
+ | 2.6556 | 86.0 | 5762 | 2.1049 |
134
+ | 2.6556 | 87.0 | 5829 | 2.0899 |
135
+ | 2.6556 | 88.0 | 5896 | 2.0796 |
136
+ | 2.6556 | 89.0 | 5963 | 2.0649 |
137
+ | 2.5131 | 90.0 | 6030 | 2.0534 |
138
+ | 2.5131 | 91.0 | 6097 | 2.0443 |
139
+ | 2.5131 | 92.0 | 6164 | 2.0360 |
140
+ | 2.5131 | 93.0 | 6231 | 2.0258 |
141
+ | 2.5131 | 94.0 | 6298 | 2.0190 |
142
+ | 2.5131 | 95.0 | 6365 | 2.0111 |
143
+ | 2.5131 | 96.0 | 6432 | 2.0100 |
144
+ | 2.5131 | 97.0 | 6499 | 2.0040 |
145
+ | 2.4077 | 98.0 | 6566 | 2.0005 |
146
+ | 2.4077 | 99.0 | 6633 | 1.9997 |
147
+ | 2.4077 | 100.0 | 6700 | 1.9992 |
148
 
149
 
150
  ### Framework versions