LiYuan199701 commited on
Commit
e0c1ce6
1 Parent(s): f999714
CESoftmaxAccuracyEvaluator_dev_results.csv ADDED
@@ -0,0 +1,379 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ epoch,steps,Accuracy
2
+ 0,100,0.345
3
+ 0,200,0.345
4
+ 0,300,0.345
5
+ 0,400,0.341
6
+ 0,500,0.439
7
+ 0,600,0.4395
8
+ 0,700,0.438
9
+ 0,800,0.44
10
+ 0,900,0.4405
11
+ 0,1000,0.447
12
+ 0,1100,0.458
13
+ 0,1200,0.4545
14
+ 0,1300,0.454
15
+ 0,1400,0.4595
16
+ 0,1500,0.46
17
+ 0,1600,0.4595
18
+ 0,1700,0.4635
19
+ 0,1800,0.4545
20
+ 0,1900,0.4625
21
+ 0,2000,0.4555
22
+ 0,2100,0.4525
23
+ 0,2200,0.46
24
+ 0,2300,0.4585
25
+ 0,2400,0.4515
26
+ 0,2500,0.45
27
+ 0,2600,0.449
28
+ 0,2700,0.4425
29
+ 0,2800,0.4545
30
+ 0,2900,0.443
31
+ 0,3000,0.441
32
+ 0,3100,0.4425
33
+ 0,3200,0.4395
34
+ 0,3300,0.4415
35
+ 0,3400,0.433
36
+ 0,3500,0.4325
37
+ 0,3600,0.38
38
+ 0,3700,0.4405
39
+ 0,3800,0.4405
40
+ 0,3900,0.4405
41
+ 0,4000,0.4405
42
+ 0,4100,0.4405
43
+ 0,4200,0.4405
44
+ 0,4300,0.4405
45
+ 0,4400,0.4405
46
+ 0,4500,0.4405
47
+ 0,4600,0.4405
48
+ 0,4700,0.4405
49
+ 0,4800,0.4405
50
+ 0,4900,0.4405
51
+ 0,5000,0.4405
52
+ 0,5100,0.4405
53
+ 0,5200,0.4405
54
+ 0,5300,0.4405
55
+ 0,5400,0.4405
56
+ 0,5500,0.4405
57
+ 0,5600,0.4405
58
+ 0,5700,0.4405
59
+ 0,5800,0.4405
60
+ 0,5900,0.4405
61
+ 0,6000,0.4405
62
+ 0,6100,0.4405
63
+ 0,6200,0.4405
64
+ 0,6300,0.4405
65
+ 0,6400,0.4405
66
+ 0,6500,0.4405
67
+ 0,6600,0.4405
68
+ 0,6700,0.4405
69
+ 0,6800,0.4405
70
+ 0,6900,0.4405
71
+ 0,7000,0.4405
72
+ 0,7100,0.4405
73
+ 0,7200,0.4405
74
+ 0,7300,0.4405
75
+ 0,7400,0.4405
76
+ 0,7500,0.4405
77
+ 0,7600,0.4405
78
+ 0,7700,0.4405
79
+ 0,7800,0.4405
80
+ 0,7900,0.4405
81
+ 0,8000,0.4405
82
+ 0,8100,0.4405
83
+ 0,8200,0.4405
84
+ 0,8300,0.4405
85
+ 0,8400,0.4405
86
+ 0,8500,0.4405
87
+ 0,8600,0.4405
88
+ 0,8700,0.4405
89
+ 0,8800,0.4405
90
+ 0,8900,0.4405
91
+ 0,9000,0.4405
92
+ 0,9100,0.4405
93
+ 0,9200,0.4405
94
+ 0,9300,0.4405
95
+ 0,9400,0.4405
96
+ 0,9500,0.4405
97
+ 0,9600,0.4405
98
+ 0,9700,0.4405
99
+ 0,9800,0.4405
100
+ 0,9900,0.4405
101
+ 0,10000,0.4405
102
+ 0,10100,0.4405
103
+ 0,10200,0.4405
104
+ 0,10300,0.4405
105
+ 0,10400,0.4405
106
+ 0,10500,0.4405
107
+ 0,10600,0.4405
108
+ 0,10700,0.4405
109
+ 0,10800,0.4405
110
+ 0,10900,0.4405
111
+ 0,11000,0.4405
112
+ 0,11100,0.4405
113
+ 0,11200,0.4405
114
+ 0,11300,0.4405
115
+ 0,11400,0.4405
116
+ 0,11500,0.4405
117
+ 0,11600,0.4405
118
+ 0,11700,0.4405
119
+ 0,11800,0.4405
120
+ 0,11900,0.4405
121
+ 0,12000,0.4405
122
+ 0,12100,0.4405
123
+ 0,12200,0.4405
124
+ 0,12300,0.4405
125
+ 0,12400,0.4405
126
+ 0,12500,0.4405
127
+ 0,-1,0.4405
128
+ 1,100,0.4405
129
+ 1,200,0.4405
130
+ 1,300,0.4405
131
+ 1,400,0.4405
132
+ 1,500,0.4405
133
+ 1,600,0.4405
134
+ 1,700,0.4405
135
+ 1,800,0.4405
136
+ 1,900,0.4405
137
+ 1,1000,0.4405
138
+ 1,1100,0.4405
139
+ 1,1200,0.4405
140
+ 1,1300,0.4405
141
+ 1,1400,0.4405
142
+ 1,1500,0.4405
143
+ 1,1600,0.4405
144
+ 1,1700,0.4405
145
+ 1,1800,0.4405
146
+ 1,1900,0.4405
147
+ 1,2000,0.4405
148
+ 1,2100,0.4405
149
+ 1,2200,0.4405
150
+ 1,2300,0.4405
151
+ 1,2400,0.4405
152
+ 1,2500,0.4405
153
+ 1,2600,0.4405
154
+ 1,2700,0.4405
155
+ 1,2800,0.4405
156
+ 1,2900,0.4405
157
+ 1,3000,0.4405
158
+ 1,3100,0.4405
159
+ 1,3200,0.4405
160
+ 1,3300,0.4405
161
+ 1,3400,0.4405
162
+ 1,3500,0.4405
163
+ 1,3600,0.4405
164
+ 1,3700,0.4405
165
+ 1,3800,0.4405
166
+ 1,3900,0.4405
167
+ 1,4000,0.4405
168
+ 1,4100,0.4405
169
+ 1,4200,0.4405
170
+ 1,4300,0.4405
171
+ 1,4400,0.4405
172
+ 1,4500,0.4405
173
+ 1,4600,0.4405
174
+ 1,4700,0.4405
175
+ 1,4800,0.4405
176
+ 1,4900,0.4405
177
+ 1,5000,0.4405
178
+ 1,5100,0.4405
179
+ 1,5200,0.4405
180
+ 1,5300,0.4405
181
+ 1,5400,0.4405
182
+ 1,5500,0.4405
183
+ 1,5600,0.4405
184
+ 1,5700,0.4405
185
+ 1,5800,0.4405
186
+ 1,5900,0.4405
187
+ 1,6000,0.4405
188
+ 1,6100,0.4405
189
+ 1,6200,0.4405
190
+ 1,6300,0.4405
191
+ 1,6400,0.4405
192
+ 1,6500,0.4405
193
+ 1,6600,0.4405
194
+ 1,6700,0.4405
195
+ 1,6800,0.4405
196
+ 1,6900,0.4405
197
+ 1,7000,0.4405
198
+ 1,7100,0.4405
199
+ 1,7200,0.4405
200
+ 1,7300,0.4405
201
+ 1,7400,0.4405
202
+ 1,7500,0.4405
203
+ 1,7600,0.4405
204
+ 1,7700,0.4405
205
+ 1,7800,0.4405
206
+ 1,7900,0.4405
207
+ 1,8000,0.4405
208
+ 1,8100,0.4405
209
+ 1,8200,0.4405
210
+ 1,8300,0.4405
211
+ 1,8400,0.4405
212
+ 1,8500,0.4405
213
+ 1,8600,0.4405
214
+ 1,8700,0.4405
215
+ 1,8800,0.4405
216
+ 1,8900,0.4405
217
+ 1,9000,0.4405
218
+ 1,9100,0.4405
219
+ 1,9200,0.4405
220
+ 1,9300,0.4405
221
+ 1,9400,0.4405
222
+ 1,9500,0.4405
223
+ 1,9600,0.4405
224
+ 1,9700,0.4405
225
+ 1,9800,0.4405
226
+ 1,9900,0.4405
227
+ 1,10000,0.4405
228
+ 1,10100,0.4405
229
+ 1,10200,0.4405
230
+ 1,10300,0.4405
231
+ 1,10400,0.4405
232
+ 1,10500,0.4405
233
+ 1,10600,0.4405
234
+ 1,10700,0.4405
235
+ 1,10800,0.4405
236
+ 1,10900,0.4405
237
+ 1,11000,0.4405
238
+ 1,11100,0.4405
239
+ 1,11200,0.4405
240
+ 1,11300,0.4405
241
+ 1,11400,0.4405
242
+ 1,11500,0.4405
243
+ 1,11600,0.4405
244
+ 1,11700,0.4405
245
+ 1,11800,0.4405
246
+ 1,11900,0.4405
247
+ 1,12000,0.4405
248
+ 1,12100,0.4405
249
+ 1,12200,0.4405
250
+ 1,12300,0.4405
251
+ 1,12400,0.4405
252
+ 1,12500,0.4405
253
+ 1,-1,0.4405
254
+ 2,100,0.4405
255
+ 2,200,0.4405
256
+ 2,300,0.4405
257
+ 2,400,0.4405
258
+ 2,500,0.4405
259
+ 2,600,0.4405
260
+ 2,700,0.4405
261
+ 2,800,0.4405
262
+ 2,900,0.4405
263
+ 2,1000,0.4405
264
+ 2,1100,0.4405
265
+ 2,1200,0.4405
266
+ 2,1300,0.4405
267
+ 2,1400,0.4405
268
+ 2,1500,0.4405
269
+ 2,1600,0.4405
270
+ 2,1700,0.4405
271
+ 2,1800,0.4405
272
+ 2,1900,0.4405
273
+ 2,2000,0.4405
274
+ 2,2100,0.4405
275
+ 2,2200,0.4405
276
+ 2,2300,0.4405
277
+ 2,2400,0.4405
278
+ 2,2500,0.4405
279
+ 2,2600,0.4405
280
+ 2,2700,0.4405
281
+ 2,2800,0.4405
282
+ 2,2900,0.4405
283
+ 2,3000,0.4405
284
+ 2,3100,0.4405
285
+ 2,3200,0.4405
286
+ 2,3300,0.4405
287
+ 2,3400,0.4405
288
+ 2,3500,0.4405
289
+ 2,3600,0.4405
290
+ 2,3700,0.4405
291
+ 2,3800,0.4405
292
+ 2,3900,0.4405
293
+ 2,4000,0.4405
294
+ 2,4100,0.4405
295
+ 2,4200,0.4405
296
+ 2,4300,0.4405
297
+ 2,4400,0.4405
298
+ 2,4500,0.4405
299
+ 2,4600,0.4405
300
+ 2,4700,0.4405
301
+ 2,4800,0.4405
302
+ 2,4900,0.4405
303
+ 2,5000,0.4405
304
+ 2,5100,0.4405
305
+ 2,5200,0.4405
306
+ 2,5300,0.4405
307
+ 2,5400,0.4405
308
+ 2,5500,0.4405
309
+ 2,5600,0.4405
310
+ 2,5700,0.4405
311
+ 2,5800,0.4405
312
+ 2,5900,0.4405
313
+ 2,6000,0.4405
314
+ 2,6100,0.4405
315
+ 2,6200,0.4405
316
+ 2,6300,0.4405
317
+ 2,6400,0.4405
318
+ 2,6500,0.4405
319
+ 2,6600,0.4405
320
+ 2,6700,0.4405
321
+ 2,6800,0.4405
322
+ 2,6900,0.4405
323
+ 2,7000,0.4405
324
+ 2,7100,0.4405
325
+ 2,7200,0.4405
326
+ 2,7300,0.4405
327
+ 2,7400,0.4405
328
+ 2,7500,0.4405
329
+ 2,7600,0.4405
330
+ 2,7700,0.4405
331
+ 2,7800,0.4405
332
+ 2,7900,0.4405
333
+ 2,8000,0.4405
334
+ 2,8100,0.4405
335
+ 2,8200,0.4405
336
+ 2,8300,0.4405
337
+ 2,8400,0.4405
338
+ 2,8500,0.4405
339
+ 2,8600,0.4405
340
+ 2,8700,0.4405
341
+ 2,8800,0.4405
342
+ 2,8900,0.4405
343
+ 2,9000,0.4405
344
+ 2,9100,0.4405
345
+ 2,9200,0.4405
346
+ 2,9300,0.4405
347
+ 2,9400,0.4405
348
+ 2,9500,0.4405
349
+ 2,9600,0.4405
350
+ 2,9700,0.4405
351
+ 2,9800,0.4405
352
+ 2,9900,0.4405
353
+ 2,10000,0.4405
354
+ 2,10100,0.4405
355
+ 2,10200,0.4405
356
+ 2,10300,0.4405
357
+ 2,10400,0.4405
358
+ 2,10500,0.4405
359
+ 2,10600,0.4405
360
+ 2,10700,0.4405
361
+ 2,10800,0.4405
362
+ 2,10900,0.4405
363
+ 2,11000,0.4405
364
+ 2,11100,0.4405
365
+ 2,11200,0.4405
366
+ 2,11300,0.4405
367
+ 2,11400,0.4405
368
+ 2,11500,0.4405
369
+ 2,11600,0.4405
370
+ 2,11700,0.4405
371
+ 2,11800,0.4405
372
+ 2,11900,0.4405
373
+ 2,12000,0.4405
374
+ 2,12100,0.4405
375
+ 2,12200,0.4405
376
+ 2,12300,0.4405
377
+ 2,12400,0.4405
378
+ 2,12500,0.4405
379
+ 2,-1,0.4405
config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "id2label": {
13
+ "0": "LABEL_0",
14
+ "1": "LABEL_1",
15
+ "2": "LABEL_2",
16
+ "3": "LABEL_3"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 3072,
20
+ "label2id": {
21
+ "LABEL_0": 0,
22
+ "LABEL_1": 1,
23
+ "LABEL_2": 2,
24
+ "LABEL_3": 3
25
+ },
26
+ "layer_norm_eps": 1e-12,
27
+ "max_position_embeddings": 512,
28
+ "model_type": "bert",
29
+ "num_attention_heads": 12,
30
+ "num_hidden_layers": 12,
31
+ "pad_token_id": 0,
32
+ "position_embedding_type": "absolute",
33
+ "torch_dtype": "float32",
34
+ "transformers_version": "4.18.0",
35
+ "type_vocab_size": 2,
36
+ "use_cache": true,
37
+ "vocab_size": 30522
38
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e149b10d3f2dce8fb0969903497c00dea3ce3f1b4628459937dce7cea237c0b
3
+ size 438012269
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-base-uncased", "tokenizer_class": "BertTokenizer"}
vocab.txt ADDED
The diff for this file is too large to render. See raw diff