dadashzadeh commited on
Commit
f1e496e
1 Parent(s): ee39dbf

End of training

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: sileod/mdeberta-v3-base-tasksource-nli
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
+ model-index:
9
+ - name: test-trainer-persian
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # test-trainer-persian
17
+
18
+ This model is a fine-tuned version of [sileod/mdeberta-v3-base-tasksource-nli](https://huggingface.co/sileod/mdeberta-v3-base-tasksource-nli) on the None dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 0.1753
21
+ - Accuracy: 0.9463
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 2e-05
41
+ - train_batch_size: 16
42
+ - eval_batch_size: 16
43
+ - seed: 42
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: linear
46
+ - num_epochs: 3
47
+
48
+ ### Training results
49
+
50
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
51
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
52
+ | 0.514 | 1.0 | 1708 | 0.3680 | 0.8829 |
53
+ | 0.3639 | 2.0 | 3416 | 0.2282 | 0.9263 |
54
+ | 0.2534 | 3.0 | 5124 | 0.1753 | 0.9463 |
55
+
56
+
57
+ ### Framework versions
58
+
59
+ - Transformers 4.35.0
60
+ - Pytorch 2.1.0+cu118
61
+ - Datasets 2.14.6
62
+ - Tokenizers 0.14.1
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[MASK]": 250101
3
+ }
config.json ADDED
@@ -0,0 +1,847 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "sileod/mdeberta-v3-base-tasksource-nli",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifiers_size": [
8
+ 3,
9
+ 3,
10
+ 3,
11
+ 3,
12
+ 3,
13
+ 3,
14
+ 3,
15
+ 3,
16
+ 3,
17
+ 3,
18
+ 3,
19
+ 3,
20
+ 3,
21
+ 3,
22
+ 3,
23
+ 3,
24
+ 3,
25
+ 1,
26
+ 1,
27
+ 1,
28
+ 1,
29
+ 1,
30
+ 1,
31
+ 1,
32
+ 1,
33
+ 1,
34
+ 1,
35
+ 2,
36
+ 2,
37
+ 2,
38
+ 2,
39
+ 2,
40
+ 2,
41
+ 2,
42
+ 31,
43
+ 12,
44
+ 31,
45
+ 11,
46
+ 15,
47
+ 2,
48
+ 2,
49
+ 2,
50
+ 2,
51
+ 2,
52
+ 2,
53
+ 6,
54
+ 6,
55
+ 6,
56
+ 2,
57
+ 8,
58
+ 10,
59
+ 2,
60
+ 2,
61
+ 5,
62
+ 2,
63
+ 2,
64
+ 2,
65
+ 2,
66
+ 1,
67
+ 1,
68
+ 1,
69
+ 20,
70
+ 235,
71
+ 1,
72
+ 1,
73
+ 1,
74
+ 1,
75
+ 1,
76
+ 1,
77
+ 1,
78
+ 1,
79
+ 1,
80
+ 1,
81
+ 1,
82
+ 1,
83
+ 1,
84
+ 1,
85
+ 1,
86
+ 1,
87
+ 1,
88
+ 1,
89
+ 1,
90
+ 1,
91
+ 1,
92
+ 1,
93
+ 1,
94
+ 1,
95
+ 1,
96
+ 1,
97
+ 1,
98
+ 1,
99
+ 1,
100
+ 1,
101
+ 1,
102
+ 1,
103
+ 1,
104
+ 1,
105
+ 1,
106
+ 1,
107
+ 1,
108
+ 1,
109
+ 1,
110
+ 1,
111
+ 1,
112
+ 1,
113
+ 1,
114
+ 1,
115
+ 1,
116
+ 1,
117
+ 1,
118
+ 1,
119
+ 1,
120
+ 1,
121
+ 1,
122
+ 1,
123
+ 1,
124
+ 1,
125
+ 1,
126
+ 1,
127
+ 1,
128
+ 1,
129
+ 1,
130
+ 1,
131
+ 1,
132
+ 1,
133
+ 1,
134
+ 1,
135
+ 1,
136
+ 9,
137
+ 17,
138
+ 18,
139
+ 18,
140
+ 18,
141
+ 18,
142
+ 18,
143
+ 18,
144
+ 18,
145
+ 18,
146
+ 18,
147
+ 18,
148
+ 18,
149
+ 18,
150
+ 18,
151
+ 18,
152
+ 18,
153
+ 18,
154
+ 18,
155
+ 18,
156
+ 18,
157
+ 18,
158
+ 18,
159
+ 18,
160
+ 18,
161
+ 18,
162
+ 18,
163
+ 18,
164
+ 18,
165
+ 18,
166
+ 18,
167
+ 18,
168
+ 18,
169
+ 18,
170
+ 18,
171
+ 18,
172
+ 18,
173
+ 18,
174
+ 18,
175
+ 18,
176
+ 18,
177
+ 18,
178
+ 18,
179
+ 18,
180
+ 18,
181
+ 18,
182
+ 18,
183
+ 18,
184
+ 18,
185
+ 18,
186
+ 18,
187
+ 18,
188
+ 18,
189
+ 18,
190
+ 18,
191
+ 18,
192
+ 18,
193
+ 18,
194
+ 18,
195
+ 18,
196
+ 18,
197
+ 18,
198
+ 18,
199
+ 18,
200
+ 18,
201
+ 18,
202
+ 18,
203
+ 18,
204
+ 18,
205
+ 18,
206
+ 18,
207
+ 18,
208
+ 18,
209
+ 18,
210
+ 18,
211
+ 18,
212
+ 18,
213
+ 18,
214
+ 18,
215
+ 18,
216
+ 18,
217
+ 18,
218
+ 18,
219
+ 18,
220
+ 18,
221
+ 18,
222
+ 18,
223
+ 18,
224
+ 18,
225
+ 18,
226
+ 18,
227
+ 18,
228
+ 18,
229
+ 18,
230
+ 18,
231
+ 18,
232
+ 18,
233
+ 18,
234
+ 18,
235
+ 18,
236
+ 18,
237
+ 18,
238
+ 18,
239
+ 18,
240
+ 18,
241
+ 18,
242
+ 18,
243
+ 18,
244
+ 18,
245
+ 18,
246
+ 18,
247
+ 18,
248
+ 18,
249
+ 18,
250
+ 18,
251
+ 18,
252
+ 18,
253
+ 18,
254
+ 18,
255
+ 18,
256
+ 18,
257
+ 18,
258
+ 18,
259
+ 18,
260
+ 18,
261
+ 18,
262
+ 18,
263
+ 18,
264
+ 18,
265
+ 18,
266
+ 18,
267
+ 18,
268
+ 18,
269
+ 18,
270
+ 18,
271
+ 18,
272
+ 18,
273
+ 18,
274
+ 18,
275
+ 18,
276
+ 18,
277
+ 18,
278
+ 18,
279
+ 18,
280
+ 18,
281
+ 18,
282
+ 18,
283
+ 18,
284
+ 18,
285
+ 18,
286
+ 18,
287
+ 18,
288
+ 18,
289
+ 18,
290
+ 18,
291
+ 18,
292
+ 18,
293
+ 18,
294
+ 18,
295
+ 18,
296
+ 18,
297
+ 18,
298
+ 18,
299
+ 18,
300
+ 18,
301
+ 18,
302
+ 18,
303
+ 18,
304
+ 18,
305
+ 18,
306
+ 18,
307
+ 18,
308
+ 18,
309
+ 18,
310
+ 18,
311
+ 18,
312
+ 18,
313
+ 18,
314
+ 18,
315
+ 18,
316
+ 18,
317
+ 1,
318
+ 3,
319
+ 3,
320
+ 3,
321
+ 3,
322
+ 3,
323
+ 3,
324
+ 3,
325
+ 3,
326
+ 3,
327
+ 3,
328
+ 6,
329
+ 5,
330
+ 3,
331
+ 1,
332
+ 60,
333
+ 60,
334
+ 60,
335
+ 60,
336
+ 60,
337
+ 60,
338
+ 60,
339
+ 60,
340
+ 60,
341
+ 60,
342
+ 60,
343
+ 60,
344
+ 60,
345
+ 60,
346
+ 60,
347
+ 60,
348
+ 60,
349
+ 60,
350
+ 60,
351
+ 60,
352
+ 60,
353
+ 60,
354
+ 60,
355
+ 60,
356
+ 60,
357
+ 60,
358
+ 60,
359
+ 60,
360
+ 60,
361
+ 60,
362
+ 60,
363
+ 60,
364
+ 60,
365
+ 60,
366
+ 60,
367
+ 60,
368
+ 60,
369
+ 60,
370
+ 60,
371
+ 60,
372
+ 60,
373
+ 60,
374
+ 60,
375
+ 60,
376
+ 60,
377
+ 60,
378
+ 60,
379
+ 60,
380
+ 60,
381
+ 60,
382
+ 60,
383
+ 60,
384
+ 60,
385
+ 60,
386
+ 2,
387
+ 73,
388
+ 73,
389
+ 73,
390
+ 73,
391
+ 73,
392
+ 73,
393
+ 73,
394
+ 73,
395
+ 73,
396
+ 73,
397
+ 73,
398
+ 73,
399
+ 73,
400
+ 113,
401
+ 3
402
+ ],
403
+ "hidden_act": "gelu",
404
+ "hidden_dropout_prob": 0.1,
405
+ "hidden_size": 768,
406
+ "id2label": {
407
+ "0": "\u06af\u0641\u062a\u06af\u0648",
408
+ "1": "\u0627\u062f\u0628\u06cc\u0627\u062a",
409
+ "2": "\u0633\u06cc\u0627\u0633\u062a",
410
+ "3": "\u0639\u0644\u0645",
411
+ "4": "\u062f\u0627\u0631\u0648",
412
+ "5": "\u0642\u0627\u0646\u0648\u0646",
413
+ "6": "\u0634\u0639\u0631",
414
+ "7": "\u062f\u06cc\u0646"
415
+ },
416
+ "initializer_range": 0.02,
417
+ "intermediate_size": 3072,
418
+ "label2id": {
419
+ "\u0627\u062f\u0628\u06cc\u0627\u062a": 1,
420
+ "\u062f\u0627\u0631\u0648": 4,
421
+ "\u062f\u06cc\u0646": 7,
422
+ "\u0633\u06cc\u0627\u0633\u062a": 2,
423
+ "\u0634\u0639\u0631": 6,
424
+ "\u0639\u0644\u0645": 3,
425
+ "\u0642\u0627\u0646\u0648\u0646": 5,
426
+ "\u06af\u0641\u062a\u06af\u0648": 0
427
+ },
428
+ "layer_norm_eps": 1e-07,
429
+ "max_position_embeddings": 512,
430
+ "max_relative_positions": -1,
431
+ "model_type": "deberta-v2",
432
+ "norm_rel_ebd": "layer_norm",
433
+ "num_attention_heads": 12,
434
+ "num_hidden_layers": 12,
435
+ "pad_token_id": 0,
436
+ "pooler_dropout": 0,
437
+ "pooler_hidden_act": "gelu",
438
+ "pooler_hidden_size": 768,
439
+ "pos_att_type": [
440
+ "p2c",
441
+ "c2p"
442
+ ],
443
+ "position_biased_input": false,
444
+ "position_buckets": 256,
445
+ "relative_attention": true,
446
+ "share_att_key": true,
447
+ "tasks": [
448
+ "multilingual-NLI-26lang-2mil7",
449
+ "xnli/ru",
450
+ "xnli/tr",
451
+ "xnli/ur",
452
+ "xnli/vi",
453
+ "xnli/zh",
454
+ "xnli/hi",
455
+ "xnli/fr",
456
+ "xnli/es",
457
+ "xnli/en",
458
+ "xnli/el",
459
+ "xnli/de",
460
+ "xnli/bg",
461
+ "xnli/ar",
462
+ "xnli/th",
463
+ "xnli/sw",
464
+ "americas_nli/all_languages",
465
+ "stsb_multi_mt/en",
466
+ "stsb_multi_mt/fr",
467
+ "stsb_multi_mt/de",
468
+ "stsb_multi_mt/es",
469
+ "stsb_multi_mt/it",
470
+ "stsb_multi_mt/nl",
471
+ "stsb_multi_mt/pl",
472
+ "stsb_multi_mt/pt",
473
+ "stsb_multi_mt/ru",
474
+ "stsb_multi_mt/zh",
475
+ "paws-x/zh",
476
+ "paws-x/ja",
477
+ "paws-x/ko",
478
+ "paws-x/en",
479
+ "paws-x/de",
480
+ "paws-x/es",
481
+ "paws-x/fr",
482
+ "miam/vm2",
483
+ "miam/maptask",
484
+ "miam/loria",
485
+ "miam/dihana",
486
+ "miam/ilisten",
487
+ "x-stance/fr",
488
+ "x-stance/de",
489
+ "offenseval_2020/da",
490
+ "offenseval_2020/tr",
491
+ "offenseval_2020/gr",
492
+ "offenseval_2020/ar",
493
+ "offenseval_dravidian/tamil",
494
+ "offenseval_dravidian/malayalam",
495
+ "offenseval_dravidian/kannada",
496
+ "xglue/qam",
497
+ "x-fact",
498
+ "xglue/nc",
499
+ "xglue/qadsm",
500
+ "xglue/qam",
501
+ "xglue/wpr",
502
+ "xlwic/xlwic_fr_fr",
503
+ "xlwic/xlwic_en_ko",
504
+ "xlwic/xlwic_it_it",
505
+ "xlwic/xlwic_de_de",
506
+ "oasst1_dense_flat/quality",
507
+ "oasst1_dense_flat/toxicity",
508
+ "oasst1_dense_flat/helpfulness",
509
+ "language-identification",
510
+ "wili_2018",
511
+ "exams/multilingual",
512
+ "xcsr/X-CSQA-ar",
513
+ "xcsr/X-CODAH-zh",
514
+ "xcsr/X-CODAH-de",
515
+ "xcsr/X-CSQA-ru",
516
+ "xcsr/X-CODAH-fr",
517
+ "xcsr/X-CODAH-it",
518
+ "xcsr/X-CODAH-jap",
519
+ "xcsr/X-CODAH-nl",
520
+ "xcsr/X-CODAH-pt",
521
+ "xcsr/X-CODAH-en",
522
+ "xcsr/X-CODAH-ru",
523
+ "xcsr/X-CODAH-ar",
524
+ "xcsr/X-CODAH-vi",
525
+ "xcsr/X-CODAH-hi",
526
+ "xcsr/X-CODAH-sw",
527
+ "xcsr/X-CODAH-ur",
528
+ "xcsr/X-CODAH-pl",
529
+ "xcsr/X-CSQA-ur",
530
+ "xcsr/X-CODAH-es",
531
+ "xcsr/X-CSQA-pt",
532
+ "xcsr/X-CSQA-vi",
533
+ "xcsr/X-CSQA-hi",
534
+ "xcsr/X-CSQA-pl",
535
+ "xcsr/X-CSQA-sw",
536
+ "xcsr/X-CSQA-nl",
537
+ "xcsr/X-CSQA-jap",
538
+ "xcsr/X-CSQA-it",
539
+ "xcsr/X-CSQA-es",
540
+ "xcsr/X-CSQA-fr",
541
+ "xcsr/X-CSQA-zh",
542
+ "xcsr/X-CSQA-en",
543
+ "xcsr/X-CSQA-de",
544
+ "xcopa/qu",
545
+ "xcopa/it",
546
+ "xcopa/ht",
547
+ "xcopa/et",
548
+ "xcopa/vi",
549
+ "xcopa/id",
550
+ "xcopa/translation-et",
551
+ "xcopa/th",
552
+ "xcopa/sw",
553
+ "xcopa/translation-sw",
554
+ "xcopa/translation-ht",
555
+ "xcopa/translation-it",
556
+ "xcopa/ta",
557
+ "xcopa/translation-zh",
558
+ "xcopa/translation-vi",
559
+ "xcopa/translation-id",
560
+ "xcopa/translation-tr",
561
+ "xcopa/translation-th",
562
+ "xcopa/translation-ta",
563
+ "xcopa/zh",
564
+ "xcopa/tr",
565
+ "xstory_cloze/eu",
566
+ "xstory_cloze/my",
567
+ "xstory_cloze/te",
568
+ "xstory_cloze/sw",
569
+ "xstory_cloze/en",
570
+ "xstory_cloze/ru",
571
+ "xstory_cloze/zh",
572
+ "xstory_cloze/es",
573
+ "xstory_cloze/ar",
574
+ "xstory_cloze/hi",
575
+ "xstory_cloze/id",
576
+ "xglue/ner",
577
+ "xglue/pos",
578
+ "universal_dependencies/sme_giella/pos",
579
+ "universal_dependencies/no_bokmaal/pos",
580
+ "universal_dependencies/no_nynorsk/pos",
581
+ "universal_dependencies/no_nynorsklia/pos",
582
+ "universal_dependencies/cu_proiel/pos",
583
+ "universal_dependencies/fro_srcmf/pos",
584
+ "universal_dependencies/orv_rnc/pos",
585
+ "universal_dependencies/pl_lfg/pos",
586
+ "universal_dependencies/otk_tonqq/pos",
587
+ "universal_dependencies/fa_perdt/pos",
588
+ "universal_dependencies/fa_seraji/pos",
589
+ "universal_dependencies/pcm_nsc/pos",
590
+ "universal_dependencies/pl_pdb/pos",
591
+ "universal_dependencies/pl_pud/pos",
592
+ "universal_dependencies/pt_bosque/pos",
593
+ "universal_dependencies/pt_gsd/pos",
594
+ "universal_dependencies/pt_pud/pos",
595
+ "universal_dependencies/orv_torot/pos",
596
+ "universal_dependencies/myu_tudet/pos",
597
+ "universal_dependencies/gv_cadhan/pos",
598
+ "universal_dependencies/gun_thomas/pos",
599
+ "universal_dependencies/koi_uh/pos",
600
+ "universal_dependencies/kpv_ikdp/pos",
601
+ "universal_dependencies/kpv_lattice/pos",
602
+ "universal_dependencies/ko_gsd/pos",
603
+ "universal_dependencies/ko_kaist/pos",
604
+ "universal_dependencies/ko_pud/pos",
605
+ "universal_dependencies/kmr_mg/pos",
606
+ "universal_dependencies/la_ittb/pos",
607
+ "universal_dependencies/la_llct/pos",
608
+ "universal_dependencies/la_perseus/pos",
609
+ "universal_dependencies/la_proiel/pos",
610
+ "universal_dependencies/lv_lvtb/pos",
611
+ "universal_dependencies/lt_alksnis/pos",
612
+ "universal_dependencies/lt_hse/pos",
613
+ "universal_dependencies/olo_kkpp/pos",
614
+ "universal_dependencies/mt_mudt/pos",
615
+ "universal_dependencies/ro_nonstandard/pos",
616
+ "universal_dependencies/mr_ufal/pos",
617
+ "universal_dependencies/gun_dooley/pos",
618
+ "universal_dependencies/mdf_jr/pos",
619
+ "universal_dependencies/ro_rrt/pos",
620
+ "universal_dependencies/ru_taiga/pos",
621
+ "universal_dependencies/ru_gsd/pos",
622
+ "universal_dependencies/ta_mwtt/pos",
623
+ "universal_dependencies/ta_ttb/pos",
624
+ "universal_dependencies/te_mtg/pos",
625
+ "universal_dependencies/th_pud/pos",
626
+ "universal_dependencies/qtd_sagt/pos",
627
+ "universal_dependencies/tr_boun/pos",
628
+ "universal_dependencies/tr_gb/pos",
629
+ "universal_dependencies/tr_imst/pos",
630
+ "universal_dependencies/tr_pud/pos",
631
+ "universal_dependencies/uk_iu/pos",
632
+ "universal_dependencies/hsb_ufal/pos",
633
+ "universal_dependencies/ur_udtb/pos",
634
+ "universal_dependencies/ug_udt/pos",
635
+ "universal_dependencies/vi_vtb/pos",
636
+ "universal_dependencies/wbp_ufal/pos",
637
+ "universal_dependencies/cy_ccg/pos",
638
+ "universal_dependencies/wo_wtb/pos",
639
+ "universal_dependencies/yo_ytb/pos",
640
+ "universal_dependencies/tl_ugnayan/pos",
641
+ "universal_dependencies/ro_simonero/pos",
642
+ "universal_dependencies/tl_trg/pos",
643
+ "universal_dependencies/sv_talbanken/pos",
644
+ "universal_dependencies/ru_pud/pos",
645
+ "universal_dependencies/ru_syntagrus/pos",
646
+ "universal_dependencies/sa_ufal/pos",
647
+ "universal_dependencies/sa_vedic/pos",
648
+ "universal_dependencies/gd_arcosg/pos",
649
+ "universal_dependencies/sr_set/pos",
650
+ "universal_dependencies/sms_giellagas/pos",
651
+ "universal_dependencies/sk_snk/pos",
652
+ "universal_dependencies/sl_ssj/pos",
653
+ "universal_dependencies/sl_sst/pos",
654
+ "universal_dependencies/ajp_madar/pos",
655
+ "universal_dependencies/es_ancora/pos",
656
+ "universal_dependencies/es_gsd/pos",
657
+ "universal_dependencies/es_pud/pos",
658
+ "universal_dependencies/swl_sslc/pos",
659
+ "universal_dependencies/sv_lines/pos",
660
+ "universal_dependencies/sv_pud/pos",
661
+ "universal_dependencies/gsw_uzh/pos",
662
+ "universal_dependencies/kk_ktb/pos",
663
+ "universal_dependencies/hi_hdtb/pos",
664
+ "universal_dependencies/ja_pud/pos",
665
+ "universal_dependencies/zh_gsd/pos",
666
+ "universal_dependencies/zh_gsdsimp/pos",
667
+ "universal_dependencies/zh_hk/pos",
668
+ "universal_dependencies/zh_pud/pos",
669
+ "universal_dependencies/ckt_hse/pos",
670
+ "universal_dependencies/lzh_kyoto/pos",
671
+ "universal_dependencies/cop_scriptorium/pos",
672
+ "universal_dependencies/hr_set/pos",
673
+ "universal_dependencies/cs_cac/pos",
674
+ "universal_dependencies/cs_cltt/pos",
675
+ "universal_dependencies/cs_fictree/pos",
676
+ "universal_dependencies/cs_pdt/pos",
677
+ "universal_dependencies/cs_pud/pos",
678
+ "universal_dependencies/da_ddt/pos",
679
+ "universal_dependencies/nl_alpino/pos",
680
+ "universal_dependencies/nl_lassysmall/pos",
681
+ "universal_dependencies/en_esl/pos",
682
+ "universal_dependencies/en_ewt/pos",
683
+ "universal_dependencies/en_gum/pos",
684
+ "universal_dependencies/zh_cfl/pos",
685
+ "universal_dependencies/ca_ancora/pos",
686
+ "universal_dependencies/yue_hk/pos",
687
+ "universal_dependencies/bxr_bdt/pos",
688
+ "universal_dependencies/af_afribooms/pos",
689
+ "universal_dependencies/krl_kkpp/pos",
690
+ "universal_dependencies/akk_riao/pos",
691
+ "universal_dependencies/aqz_tudet/pos",
692
+ "universal_dependencies/sq_tsa/pos",
693
+ "universal_dependencies/am_att/pos",
694
+ "universal_dependencies/grc_perseus/pos",
695
+ "universal_dependencies/grc_proiel/pos",
696
+ "universal_dependencies/apu_ufpa/pos",
697
+ "universal_dependencies/en_gumreddit/pos",
698
+ "universal_dependencies/ar_nyuad/pos",
699
+ "universal_dependencies/ar_pud/pos",
700
+ "universal_dependencies/hy_armtdp/pos",
701
+ "universal_dependencies/aii_as/pos",
702
+ "universal_dependencies/bm_crb/pos",
703
+ "universal_dependencies/eu_bdt/pos",
704
+ "universal_dependencies/be_hse/pos",
705
+ "universal_dependencies/bho_bhtb/pos",
706
+ "universal_dependencies/br_keb/pos",
707
+ "universal_dependencies/bg_btb/pos",
708
+ "universal_dependencies/ar_padt/pos",
709
+ "universal_dependencies/en_lines/pos",
710
+ "universal_dependencies/akk_pisandub/pos",
711
+ "universal_dependencies/en_pronouns/pos",
712
+ "universal_dependencies/el_gdt/pos",
713
+ "universal_dependencies/he_htb/pos",
714
+ "universal_dependencies/qhe_hiencs/pos",
715
+ "universal_dependencies/hi_pud/pos",
716
+ "universal_dependencies/hu_szeged/pos",
717
+ "universal_dependencies/is_icepahc/pos",
718
+ "universal_dependencies/id_csui/pos",
719
+ "universal_dependencies/id_gsd/pos",
720
+ "universal_dependencies/id_pud/pos",
721
+ "universal_dependencies/ga_idt/pos",
722
+ "universal_dependencies/it_isdt/pos",
723
+ "universal_dependencies/it_partut/pos",
724
+ "universal_dependencies/it_postwita/pos",
725
+ "universal_dependencies/it_pud/pos",
726
+ "universal_dependencies/it_twittiro/pos",
727
+ "universal_dependencies/it_vit/pos",
728
+ "universal_dependencies/ja_bccwj/pos",
729
+ "universal_dependencies/ja_gsd/pos",
730
+ "universal_dependencies/ja_modern/pos",
731
+ "universal_dependencies/got_proiel/pos",
732
+ "universal_dependencies/de_pud/pos",
733
+ "universal_dependencies/is_pud/pos",
734
+ "universal_dependencies/de_hdt/pos",
735
+ "universal_dependencies/en_pud/pos",
736
+ "universal_dependencies/myv_jr/pos",
737
+ "universal_dependencies/de_lit/pos",
738
+ "universal_dependencies/et_ewt/pos",
739
+ "universal_dependencies/fo_farpahc/pos",
740
+ "universal_dependencies/fo_oft/pos",
741
+ "universal_dependencies/fi_ftb/pos",
742
+ "universal_dependencies/fi_ood/pos",
743
+ "universal_dependencies/fi_pud/pos",
744
+ "universal_dependencies/fi_tdt/pos",
745
+ "universal_dependencies/et_edt/pos",
746
+ "universal_dependencies/fr_ftb/pos",
747
+ "universal_dependencies/fr_fqb/pos",
748
+ "universal_dependencies/de_gsd/pos",
749
+ "universal_dependencies/gl_treegal/pos",
750
+ "universal_dependencies/gl_ctg/pos",
751
+ "universal_dependencies/fr_spoken/pos",
752
+ "universal_dependencies/en_partut/pos",
753
+ "universal_dependencies/fr_pud/pos",
754
+ "universal_dependencies/fr_partut/pos",
755
+ "universal_dependencies/fr_sequoia/pos",
756
+ "universal_dependencies/fr_gsd/pos",
757
+ "oasst1_pairwise_rlhf_reward",
758
+ "multilingual-sentiments/all",
759
+ "tweet_sentiment_multilingual/arabic",
760
+ "tweet_sentiment_multilingual/french",
761
+ "tweet_sentiment_multilingual/english",
762
+ "tweet_sentiment_multilingual/hindi",
763
+ "tweet_sentiment_multilingual/portuguese",
764
+ "tweet_sentiment_multilingual/spanish",
765
+ "tweet_sentiment_multilingual/all",
766
+ "tweet_sentiment_multilingual/german",
767
+ "tweet_sentiment_multilingual/italian",
768
+ "amazon_reviews_multi/all_languages",
769
+ "universal-joy",
770
+ "mms",
771
+ "ACES",
772
+ "massive/my-MM",
773
+ "massive/ro-RO",
774
+ "massive/pt-PT",
775
+ "massive/pl-PL",
776
+ "massive/nl-NL",
777
+ "massive/nb-NO",
778
+ "massive/es-ES",
779
+ "massive/ms-MY",
780
+ "massive/mn-MN",
781
+ "massive/ml-IN",
782
+ "massive/lv-LV",
783
+ "massive/ko-KR",
784
+ "massive/ru-RU",
785
+ "massive/kn-IN",
786
+ "massive/ka-GE",
787
+ "massive/jv-ID",
788
+ "massive/ja-JP",
789
+ "massive/it-IT",
790
+ "massive/is-IS",
791
+ "massive/id-ID",
792
+ "massive/hy-AM",
793
+ "massive/hu-HU",
794
+ "massive/hi-IN",
795
+ "massive/he-IL",
796
+ "massive/fr-FR",
797
+ "massive/km-KH",
798
+ "massive/fi-FI",
799
+ "massive/sl-SL",
800
+ "massive/sv-SE",
801
+ "massive/af-ZA",
802
+ "massive/am-ET",
803
+ "massive/ar-SA",
804
+ "massive/az-AZ",
805
+ "massive/bn-BD",
806
+ "massive/ca-ES",
807
+ "massive/cy-GB",
808
+ "massive/da-DK",
809
+ "massive/de-DE",
810
+ "massive/el-GR",
811
+ "massive/sq-AL",
812
+ "massive/en-US",
813
+ "massive/all",
814
+ "massive/zh-TW",
815
+ "massive/zh-CN",
816
+ "massive/vi-VN",
817
+ "massive/ur-PK",
818
+ "massive/tr-TR",
819
+ "massive/tl-PH",
820
+ "massive/th-TH",
821
+ "massive/te-IN",
822
+ "massive/ta-IN",
823
+ "massive/sw-KE",
824
+ "massive/all_1.1",
825
+ "massive/fa-IR",
826
+ "tydi-as2-balanced",
827
+ "multiconer_v2/Hindi (HI)",
828
+ "multiconer_v2/Multilingual (MULTI)",
829
+ "multiconer_v2/Ukrainian (UK)",
830
+ "multiconer_v2/Swedish (SV)",
831
+ "multiconer_v2/Spanish (ES)",
832
+ "multiconer_v2/Bangla (BN)",
833
+ "multiconer_v2/Chinese (ZH)",
834
+ "multiconer_v2/English (EN)",
835
+ "multiconer_v2/Farsi (FA)",
836
+ "multiconer_v2/Portuguese (PT)",
837
+ "multiconer_v2/German (DE)",
838
+ "multiconer_v2/Italian (IT)",
839
+ "multiconer_v2/French (FR)",
840
+ "mtop",
841
+ "multilingual-zero-shot-label-nli"
842
+ ],
843
+ "torch_dtype": "float32",
844
+ "transformers_version": "4.35.0",
845
+ "type_vocab_size": 0,
846
+ "vocab_size": 251000
847
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f642f9884d77f61d21885e4c5e231c7b5d9758088dda771fec4d7883837bb316
3
+ size 1115286664
runs/Nov13_19-31-31_a3eb61d85f3f/events.out.tfevents.1699903907.a3eb61d85f3f.239.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cce743095f1e8a87eeaddaf61d290faca63c7e6575583b68f136640fe7cf5be
3
+ size 23746
runs/Nov13_19-31-31_a3eb61d85f3f/events.out.tfevents.1699906866.a3eb61d85f3f.239.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51c300b007fded700a0d7f4e577ba23472e52a055456acbf1b7f8dba3820e444
3
+ size 411
special_tokens_map.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": "[UNK]"
9
+ }
spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13c8d666d62a7bc4ac8f040aab68e942c861f93303156cc28f5c7e885d86d6e3
3
+ size 4305025
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f27bde916567fb4107bb534811fc793175a893c3de63115dc7e8598d4aaf6fd2
3
+ size 16316081
tokenizer_config.json ADDED
@@ -0,0 +1,65 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "250101": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_lower_case": false,
48
+ "eos_token": "[SEP]",
49
+ "mask_token": "[MASK]",
50
+ "max_length": 256,
51
+ "model_max_length": 1000000000000000019884624838656,
52
+ "pad_to_multiple_of": null,
53
+ "pad_token": "[PAD]",
54
+ "pad_token_type_id": 0,
55
+ "padding_side": "right",
56
+ "sep_token": "[SEP]",
57
+ "sp_model_kwargs": {},
58
+ "split_by_punct": false,
59
+ "stride": 0,
60
+ "tokenizer_class": "DebertaV2Tokenizer",
61
+ "truncation_side": "right",
62
+ "truncation_strategy": "longest_first",
63
+ "unk_token": "[UNK]",
64
+ "vocab_type": "spm"
65
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d61078321ca9e5741bc65f70c07488e68d2348091e0def579c94523b63253a6e
3
+ size 4600