Zero-Shot Classification
Transformers
PyTorch
Safetensors
English
deberta-v2
text-classification
deberta-v3-base
deberta-v3
deberta
nli
natural-language-inference
multitask
multi-task
pipeline
extreme-multi-task
extreme-mtl
tasksource
zero-shot
rlhf
Eval Results
Inference Endpoints
sileod commited on
Commit
fac6592
1 Parent(s): 5bcc815

Upload DebertaV2ForSequenceClassification

Browse files
Files changed (2) hide show
  1. config.json +651 -2045
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -1,1802 +1,439 @@
1
  {
2
- "_name_or_path": "deberta-tasksource-full",
3
  "architectures": [
4
  "DebertaV2ForSequenceClassification"
5
  ],
6
- "id2label": {
7
- "0": "entailment",
8
- "1": "neutral",
9
- "2": "contradiction"
10
- },
11
- "label2id": {
12
- "entailment": 0,
13
- "neutral": 1,
14
- "contradiction": 2
15
- },
16
  "attention_probs_dropout_prob": 0.1,
17
  "classifiers_size": [
18
- [
19
- 2,
20
- 768
21
- ],
22
- [
23
- 2,
24
- 768
25
- ],
26
- [
27
- 2,
28
- 768
29
- ],
30
- [
31
- 2,
32
- 768
33
- ],
34
- [
35
- 2,
36
- 768
37
- ],
38
- [
39
- 2,
40
- 768
41
- ],
42
- [
43
- 2,
44
- 768
45
- ],
46
- [
47
- 2,
48
- 768
49
- ],
50
- [
51
- 2,
52
- 768
53
- ],
54
- [
55
- 2,
56
- 768
57
- ],
58
- [
59
- 2,
60
- 768
61
- ],
62
- [
63
- 2,
64
- 768
65
- ],
66
- [
67
- 2,
68
- 768
69
- ],
70
- [
71
- 2,
72
- 768
73
- ],
74
- [
75
- 2,
76
- 768
77
- ],
78
- [
79
- 2,
80
- 768
81
- ],
82
- [
83
- 2,
84
- 768
85
- ],
86
- [
87
- 2,
88
- 768
89
- ],
90
- [
91
- 2,
92
- 768
93
- ],
94
- [
95
- 3,
96
- 768
97
- ],
98
- [
99
- 3,
100
- 768
101
- ],
102
- [
103
- 3,
104
- 768
105
- ],
106
- [
107
- 3,
108
- 768
109
- ],
110
- [
111
- 1,
112
- 768
113
- ],
114
- [
115
- 3,
116
- 768
117
- ],
118
- [
119
- 3,
120
- 768
121
- ],
122
- [
123
- 3,
124
- 768
125
- ],
126
- [
127
- 2,
128
- 768
129
- ],
130
- [
131
- 2,
132
- 768
133
- ],
134
- [
135
- 3,
136
- 768
137
- ],
138
- [
139
- 6,
140
- 768
141
- ],
142
- [
143
- 2,
144
- 768
145
- ],
146
- [
147
- 2,
148
- 768
149
- ],
150
- [
151
- 2,
152
- 768
153
- ],
154
- [
155
- 2,
156
- 768
157
- ],
158
- [
159
- 2,
160
- 768
161
- ],
162
- [
163
- 2,
164
- 768
165
- ],
166
- [
167
- 2,
168
- 768
169
- ],
170
- [
171
- 2,
172
- 768
173
- ],
174
- [
175
- 2,
176
- 768
177
- ],
178
- [
179
- 2,
180
- 768
181
- ],
182
- [
183
- 3,
184
- 768
185
- ],
186
- [
187
- 3,
188
- 768
189
- ],
190
- [
191
- 3,
192
- 768
193
- ],
194
- [
195
- 3,
196
- 768
197
- ],
198
- [
199
- 3,
200
- 768
201
- ],
202
- [
203
- 3,
204
- 768
205
- ],
206
- [
207
- 3,
208
- 768
209
- ],
210
- [
211
- 2,
212
- 768
213
- ],
214
- [
215
- 3,
216
- 768
217
- ],
218
- [
219
- 2,
220
- 768
221
- ],
222
- [
223
- 2,
224
- 768
225
- ],
226
- [
227
- 2,
228
- 768
229
- ],
230
- [
231
- 3,
232
- 768
233
- ],
234
- [
235
- 3,
236
- 768
237
- ],
238
- [
239
- 3,
240
- 768
241
- ],
242
- [
243
- 3,
244
- 768
245
- ],
246
- [
247
- 3,
248
- 768
249
- ],
250
- [
251
- 3,
252
- 768
253
- ],
254
- [
255
- 3,
256
- 768
257
- ],
258
- [
259
- 3,
260
- 768
261
- ],
262
- [
263
- 3,
264
- 768
265
- ],
266
- [
267
- 3,
268
- 768
269
- ],
270
- [
271
- 2,
272
- 768
273
- ],
274
- [
275
- 2,
276
- 768
277
- ],
278
- [
279
- 3,
280
- 768
281
- ],
282
- [
283
- 3,
284
- 768
285
- ],
286
- [
287
- 3,
288
- 768
289
- ],
290
- [
291
- 3,
292
- 768
293
- ],
294
- [
295
- 3,
296
- 768
297
- ],
298
- [
299
- 3,
300
- 768
301
- ],
302
- [
303
- 2,
304
- 768
305
- ],
306
- [
307
- 3,
308
- 768
309
- ],
310
- [
311
- 3,
312
- 768
313
- ],
314
- [
315
- 3,
316
- 768
317
- ],
318
- [
319
- 3,
320
- 768
321
- ],
322
- [
323
- 3,
324
- 768
325
- ],
326
- [
327
- 3,
328
- 768
329
- ],
330
- [
331
- 3,
332
- 768
333
- ],
334
- [
335
- 3,
336
- 768
337
- ],
338
- [
339
- 3,
340
- 768
341
- ],
342
- [
343
- 3,
344
- 768
345
- ],
346
- [
347
- 3,
348
- 768
349
- ],
350
- [
351
- 3,
352
- 768
353
- ],
354
- [
355
- 3,
356
- 768
357
- ],
358
- [
359
- 3,
360
- 768
361
- ],
362
- [
363
- 3,
364
- 768
365
- ],
366
- [
367
- 3,
368
- 768
369
- ],
370
- [
371
- 3,
372
- 768
373
- ],
374
- [
375
- 3,
376
- 768
377
- ],
378
- [
379
- 3,
380
- 768
381
- ],
382
- [
383
- 3,
384
- 768
385
- ],
386
- [
387
- 3,
388
- 768
389
- ],
390
- [
391
- 3,
392
- 768
393
- ],
394
- [
395
- 3,
396
- 768
397
- ],
398
- [
399
- 3,
400
- 768
401
- ],
402
- [
403
- 2,
404
- 768
405
- ],
406
- [
407
- 2,
408
- 768
409
- ],
410
- [
411
- 2,
412
- 768
413
- ],
414
- [
415
- 2,
416
- 768
417
- ],
418
- [
419
- 2,
420
- 768
421
- ],
422
- [
423
- 47,
424
- 768
425
- ],
426
- [
427
- 23,
428
- 768
429
- ],
430
- [
431
- 9,
432
- 768
433
- ],
434
- [
435
- 1,
436
- 768
437
- ],
438
- [
439
- 1,
440
- 768
441
- ],
442
- [
443
- 1,
444
- 768
445
- ],
446
- [
447
- 1,
448
- 768
449
- ],
450
- [
451
- 1,
452
- 768
453
- ],
454
- [
455
- 1,
456
- 768
457
- ],
458
- [
459
- 1,
460
- 768
461
- ],
462
- [
463
- 1,
464
- 768
465
- ],
466
- [
467
- 1,
468
- 768
469
- ],
470
- [
471
- 1,
472
- 768
473
- ],
474
- [
475
- 1,
476
- 768
477
- ],
478
- [
479
- 1,
480
- 768
481
- ],
482
- [
483
- 1,
484
- 768
485
- ],
486
- [
487
- 1,
488
- 768
489
- ],
490
- [
491
- 1,
492
- 768
493
- ],
494
- [
495
- 1,
496
- 768
497
- ],
498
- [
499
- 1,
500
- 768
501
- ],
502
- [
503
- 1,
504
- 768
505
- ],
506
- [
507
- 1,
508
- 768
509
- ],
510
- [
511
- 1,
512
- 768
513
- ],
514
- [
515
- 1,
516
- 768
517
- ],
518
- [
519
- 1,
520
- 768
521
- ],
522
- [
523
- 1,
524
- 768
525
- ],
526
- [
527
- 1,
528
- 768
529
- ],
530
- [
531
- 1,
532
- 768
533
- ],
534
- [
535
- 1,
536
- 768
537
- ],
538
- [
539
- 1,
540
- 768
541
- ],
542
- [
543
- 1,
544
- 768
545
- ],
546
- [
547
- 1,
548
- 768
549
- ],
550
- [
551
- 1,
552
- 768
553
- ],
554
- [
555
- 1,
556
- 768
557
- ],
558
- [
559
- 1,
560
- 768
561
- ],
562
- [
563
- 1,
564
- 768
565
- ],
566
- [
567
- 1,
568
- 768
569
- ],
570
- [
571
- 1,
572
- 768
573
- ],
574
- [
575
- 1,
576
- 768
577
- ],
578
- [
579
- 1,
580
- 768
581
- ],
582
- [
583
- 1,
584
- 768
585
- ],
586
- [
587
- 1,
588
- 768
589
- ],
590
- [
591
- 1,
592
- 768
593
- ],
594
- [
595
- 1,
596
- 768
597
- ],
598
- [
599
- 1,
600
- 768
601
- ],
602
- [
603
- 1,
604
- 768
605
- ],
606
- [
607
- 1,
608
- 768
609
- ],
610
- [
611
- 1,
612
- 768
613
- ],
614
- [
615
- 1,
616
- 768
617
- ],
618
- [
619
- 1,
620
- 768
621
- ],
622
- [
623
- 1,
624
- 768
625
- ],
626
- [
627
- 1,
628
- 768
629
- ],
630
- [
631
- 1,
632
- 768
633
- ],
634
- [
635
- 1,
636
- 768
637
- ],
638
- [
639
- 1,
640
- 768
641
- ],
642
- [
643
- 1,
644
- 768
645
- ],
646
- [
647
- 1,
648
- 768
649
- ],
650
- [
651
- 1,
652
- 768
653
- ],
654
- [
655
- 1,
656
- 768
657
- ],
658
- [
659
- 1,
660
- 768
661
- ],
662
- [
663
- 1,
664
- 768
665
- ],
666
- [
667
- 1,
668
- 768
669
- ],
670
- [
671
- 1,
672
- 768
673
- ],
674
- [
675
- 1,
676
- 768
677
- ],
678
- [
679
- 1,
680
- 768
681
- ],
682
- [
683
- 1,
684
- 768
685
- ],
686
- [
687
- 1,
688
- 768
689
- ],
690
- [
691
- 1,
692
- 768
693
- ],
694
- [
695
- 1,
696
- 768
697
- ],
698
- [
699
- 1,
700
- 768
701
- ],
702
- [
703
- 1,
704
- 768
705
- ],
706
- [
707
- 1,
708
- 768
709
- ],
710
- [
711
- 1,
712
- 768
713
- ],
714
- [
715
- 1,
716
- 768
717
- ],
718
- [
719
- 1,
720
- 768
721
- ],
722
- [
723
- 1,
724
- 768
725
- ],
726
- [
727
- 1,
728
- 768
729
- ],
730
- [
731
- 1,
732
- 768
733
- ],
734
- [
735
- 1,
736
- 768
737
- ],
738
- [
739
- 1,
740
- 768
741
- ],
742
- [
743
- 1,
744
- 768
745
- ],
746
- [
747
- 1,
748
- 768
749
- ],
750
- [
751
- 1,
752
- 768
753
- ],
754
- [
755
- 1,
756
- 768
757
- ],
758
- [
759
- 1,
760
- 768
761
- ],
762
- [
763
- 1,
764
- 768
765
- ],
766
- [
767
- 1,
768
- 768
769
- ],
770
- [
771
- 1,
772
- 768
773
- ],
774
- [
775
- 1,
776
- 768
777
- ],
778
- [
779
- 1,
780
- 768
781
- ],
782
- [
783
- 1,
784
- 768
785
- ],
786
- [
787
- 1,
788
- 768
789
- ],
790
- [
791
- 1,
792
- 768
793
- ],
794
- [
795
- 1,
796
- 768
797
- ],
798
- [
799
- 1,
800
- 768
801
- ],
802
- [
803
- 1,
804
- 768
805
- ],
806
- [
807
- 1,
808
- 768
809
- ],
810
- [
811
- 1,
812
- 768
813
- ],
814
- [
815
- 1,
816
- 768
817
- ],
818
- [
819
- 1,
820
- 768
821
- ],
822
- [
823
- 1,
824
- 768
825
- ],
826
- [
827
- 1,
828
- 768
829
- ],
830
- [
831
- 1,
832
- 768
833
- ],
834
- [
835
- 1,
836
- 768
837
- ],
838
- [
839
- 1,
840
- 768
841
- ],
842
- [
843
- 1,
844
- 768
845
- ],
846
- [
847
- 1,
848
- 768
849
- ],
850
- [
851
- 1,
852
- 768
853
- ],
854
- [
855
- 1,
856
- 768
857
- ],
858
- [
859
- 1,
860
- 768
861
- ],
862
- [
863
- 1,
864
- 768
865
- ],
866
- [
867
- 1,
868
- 768
869
- ],
870
- [
871
- 1,
872
- 768
873
- ],
874
- [
875
- 1,
876
- 768
877
- ],
878
- [
879
- 1,
880
- 768
881
- ],
882
- [
883
- 1,
884
- 768
885
- ],
886
- [
887
- 1,
888
- 768
889
- ],
890
- [
891
- 1,
892
- 768
893
- ],
894
- [
895
- 1,
896
- 768
897
- ],
898
- [
899
- 1,
900
- 768
901
- ],
902
- [
903
- 1,
904
- 768
905
- ],
906
- [
907
- 1,
908
- 768
909
- ],
910
- [
911
- 1,
912
- 768
913
- ],
914
- [
915
- 1,
916
- 768
917
- ],
918
- [
919
- 1,
920
- 768
921
- ],
922
- [
923
- 1,
924
- 768
925
- ],
926
- [
927
- 1,
928
- 768
929
- ],
930
- [
931
- 1,
932
- 768
933
- ],
934
- [
935
- 1,
936
- 768
937
- ],
938
- [
939
- 1,
940
- 768
941
- ],
942
- [
943
- 1,
944
- 768
945
- ],
946
- [
947
- 1,
948
- 768
949
- ],
950
- [
951
- 1,
952
- 768
953
- ],
954
- [
955
- 1,
956
- 768
957
- ],
958
- [
959
- 1,
960
- 768
961
- ],
962
- [
963
- 1,
964
- 768
965
- ],
966
- [
967
- 1,
968
- 768
969
- ],
970
- [
971
- 1,
972
- 768
973
- ],
974
- [
975
- 1,
976
- 768
977
- ],
978
- [
979
- 1,
980
- 768
981
- ],
982
- [
983
- 1,
984
- 768
985
- ],
986
- [
987
- 1,
988
- 768
989
- ],
990
- [
991
- 1,
992
- 768
993
- ],
994
- [
995
- 1,
996
- 768
997
- ],
998
- [
999
- 1,
1000
- 768
1001
- ],
1002
- [
1003
- 1,
1004
- 768
1005
- ],
1006
- [
1007
- 1,
1008
- 768
1009
- ],
1010
- [
1011
- 1,
1012
- 768
1013
- ],
1014
- [
1015
- 1,
1016
- 768
1017
- ],
1018
- [
1019
- 1,
1020
- 768
1021
- ],
1022
- [
1023
- 1,
1024
- 768
1025
- ],
1026
- [
1027
- 1,
1028
- 768
1029
- ],
1030
- [
1031
- 1,
1032
- 768
1033
- ],
1034
- [
1035
- 1,
1036
- 768
1037
- ],
1038
- [
1039
- 1,
1040
- 768
1041
- ],
1042
- [
1043
- 1,
1044
- 768
1045
- ],
1046
- [
1047
- 1,
1048
- 768
1049
- ],
1050
- [
1051
- 1,
1052
- 768
1053
- ],
1054
- [
1055
- 1,
1056
- 768
1057
- ],
1058
- [
1059
- 1,
1060
- 768
1061
- ],
1062
- [
1063
- 1,
1064
- 768
1065
- ],
1066
- [
1067
- 1,
1068
- 768
1069
- ],
1070
- [
1071
- 1,
1072
- 768
1073
- ],
1074
- [
1075
- 1,
1076
- 768
1077
- ],
1078
- [
1079
- 1,
1080
- 768
1081
- ],
1082
- [
1083
- 1,
1084
- 768
1085
- ],
1086
- [
1087
- 1,
1088
- 768
1089
- ],
1090
- [
1091
- 1,
1092
- 768
1093
- ],
1094
- [
1095
- 1,
1096
- 768
1097
- ],
1098
- [
1099
- 1,
1100
- 768
1101
- ],
1102
- [
1103
- 1,
1104
- 768
1105
- ],
1106
- [
1107
- 1,
1108
- 768
1109
- ],
1110
- [
1111
- 1,
1112
- 768
1113
- ],
1114
- [
1115
- 1,
1116
- 768
1117
- ],
1118
- [
1119
- 1,
1120
- 768
1121
- ],
1122
- [
1123
- 1,
1124
- 768
1125
- ],
1126
- [
1127
- 1,
1128
- 768
1129
- ],
1130
- [
1131
- 1,
1132
- 768
1133
- ],
1134
- [
1135
- 1,
1136
- 768
1137
- ],
1138
- [
1139
- 1,
1140
- 768
1141
- ],
1142
- [
1143
- 1,
1144
- 768
1145
- ],
1146
- [
1147
- 1,
1148
- 768
1149
- ],
1150
- [
1151
- 1,
1152
- 768
1153
- ],
1154
- [
1155
- 1,
1156
- 768
1157
- ],
1158
- [
1159
- 1,
1160
- 768
1161
- ],
1162
- [
1163
- 1,
1164
- 768
1165
- ],
1166
- [
1167
- 1,
1168
- 768
1169
- ],
1170
- [
1171
- 1,
1172
- 768
1173
- ],
1174
- [
1175
- 1,
1176
- 768
1177
- ],
1178
- [
1179
- 2,
1180
- 768
1181
- ],
1182
- [
1183
- 20,
1184
- 768
1185
- ],
1186
- [
1187
- 50,
1188
- 768
1189
- ],
1190
- [
1191
- 3,
1192
- 768
1193
- ],
1194
- [
1195
- 3,
1196
- 768
1197
- ],
1198
- [
1199
- 4,
1200
- 768
1201
- ],
1202
- [
1203
- 2,
1204
- 768
1205
- ],
1206
- [
1207
- 8,
1208
- 768
1209
- ],
1210
- [
1211
- 2,
1212
- 768
1213
- ],
1214
- [
1215
- 2,
1216
- 768
1217
- ],
1218
- [
1219
- 2,
1220
- 768
1221
- ],
1222
- [
1223
- 2,
1224
- 768
1225
- ],
1226
- [
1227
- 1,
1228
- 768
1229
- ],
1230
- [
1231
- 3,
1232
- 768
1233
- ],
1234
- [
1235
- 2,
1236
- 768
1237
- ],
1238
- [
1239
- 2,
1240
- 768
1241
- ],
1242
- [
1243
- 2,
1244
- 768
1245
- ],
1246
- [
1247
- 2,
1248
- 768
1249
- ],
1250
- [
1251
- 3,
1252
- 768
1253
- ],
1254
- [
1255
- 2,
1256
- 768
1257
- ],
1258
- [
1259
- 2,
1260
- 768
1261
- ],
1262
- [
1263
- 2,
1264
- 768
1265
- ],
1266
- [
1267
- 3,
1268
- 768
1269
- ],
1270
- [
1271
- 3,
1272
- 768
1273
- ],
1274
- [
1275
- 3,
1276
- 768
1277
- ],
1278
- [
1279
- 3,
1280
- 768
1281
- ],
1282
- [
1283
- 3,
1284
- 768
1285
- ],
1286
- [
1287
- 2,
1288
- 768
1289
- ],
1290
- [
1291
- 3,
1292
- 768
1293
- ],
1294
- [
1295
- 2,
1296
- 768
1297
- ],
1298
- [
1299
- 4,
1300
- 768
1301
- ],
1302
- [
1303
- 20,
1304
- 768
1305
- ],
1306
- [
1307
- 2,
1308
- 768
1309
- ],
1310
- [
1311
- 174,
1312
- 768
1313
- ],
1314
- [
1315
- 41,
1316
- 768
1317
- ],
1318
- [
1319
- 2,
1320
- 768
1321
- ],
1322
- [
1323
- 2,
1324
- 768
1325
- ],
1326
- [
1327
- 2,
1328
- 768
1329
- ],
1330
- [
1331
- 2,
1332
- 768
1333
- ],
1334
- [
1335
- 51,
1336
- 768
1337
- ],
1338
- [
1339
- 3,
1340
- 768
1341
- ],
1342
- [
1343
- 2,
1344
- 768
1345
- ],
1346
- [
1347
- 2,
1348
- 768
1349
- ],
1350
- [
1351
- 17,
1352
- 768
1353
- ],
1354
- [
1355
- 3,
1356
- 768
1357
- ],
1358
- [
1359
- 8,
1360
- 768
1361
- ],
1362
- [
1363
- 16,
1364
- 768
1365
- ],
1366
- [
1367
- 2,
1368
- 768
1369
- ],
1370
- [
1371
- 2,
1372
- 768
1373
- ],
1374
- [
1375
- 2,
1376
- 768
1377
- ],
1378
- [
1379
- 2,
1380
- 768
1381
- ],
1382
- [
1383
- 18,
1384
- 768
1385
- ],
1386
- [
1387
- 3,
1388
- 768
1389
- ],
1390
- [
1391
- 2,
1392
- 768
1393
- ],
1394
- [
1395
- 100,
1396
- 768
1397
- ],
1398
- [
1399
- 13,
1400
- 768
1401
- ],
1402
- [
1403
- 100,
1404
- 768
1405
- ],
1406
- [
1407
- 8,
1408
- 768
1409
- ],
1410
- [
1411
- 1,
1412
- 768
1413
- ],
1414
- [
1415
- 2,
1416
- 768
1417
- ],
1418
- [
1419
- 2,
1420
- 768
1421
- ],
1422
- [
1423
- 4,
1424
- 768
1425
- ],
1426
- [
1427
- 5,
1428
- 768
1429
- ],
1430
- [
1431
- 3,
1432
- 768
1433
- ],
1434
- [
1435
- 4,
1436
- 768
1437
- ],
1438
- [
1439
- 14,
1440
- 768
1441
- ],
1442
- [
1443
- 2,
1444
- 768
1445
- ],
1446
- [
1447
- 6,
1448
- 768
1449
- ],
1450
- [
1451
- 4,
1452
- 768
1453
- ],
1454
- [
1455
- 2,
1456
- 768
1457
- ],
1458
- [
1459
- 1,
1460
- 768
1461
- ],
1462
- [
1463
- 3,
1464
- 768
1465
- ],
1466
- [
1467
- 10,
1468
- 768
1469
- ],
1470
- [
1471
- 77,
1472
- 768
1473
- ],
1474
- [
1475
- 3,
1476
- 768
1477
- ],
1478
- [
1479
- 2,
1480
- 768
1481
- ],
1482
- [
1483
- 2,
1484
- 768
1485
- ],
1486
- [
1487
- 28,
1488
- 768
1489
- ],
1490
- [
1491
- 3,
1492
- 768
1493
- ],
1494
- [
1495
- 6,
1496
- 768
1497
- ],
1498
- [
1499
- 3,
1500
- 768
1501
- ],
1502
- [
1503
- 7,
1504
- 768
1505
- ],
1506
- [
1507
- 5,
1508
- 768
1509
- ],
1510
- [
1511
- 6,
1512
- 768
1513
- ],
1514
- [
1515
- 4,
1516
- 768
1517
- ],
1518
- [
1519
- 2,
1520
- 768
1521
- ],
1522
- [
1523
- 2,
1524
- 768
1525
- ],
1526
- [
1527
- 1000,
1528
- 768
1529
- ],
1530
- [
1531
- 2,
1532
- 768
1533
- ],
1534
- [
1535
- 2,
1536
- 768
1537
- ],
1538
- [
1539
- 7,
1540
- 768
1541
- ],
1542
- [
1543
- 6,
1544
- 768
1545
- ],
1546
- [
1547
- 20,
1548
- 768
1549
- ],
1550
- [
1551
- 2,
1552
- 768
1553
- ],
1554
- [
1555
- 2,
1556
- 768
1557
- ],
1558
- [
1559
- 4,
1560
- 768
1561
- ],
1562
- [
1563
- 2,
1564
- 768
1565
- ],
1566
- [
1567
- 3,
1568
- 768
1569
- ],
1570
- [
1571
- 3,
1572
- 768
1573
- ],
1574
- [
1575
- 2,
1576
- 768
1577
- ],
1578
- [
1579
- 2,
1580
- 768
1581
- ],
1582
- [
1583
- 9,
1584
- 768
1585
- ],
1586
- [
1587
- 13,
1588
- 768
1589
- ],
1590
- [
1591
- 4,
1592
- 768
1593
- ],
1594
- [
1595
- 2,
1596
- 768
1597
- ],
1598
- [
1599
- 2,
1600
- 768
1601
- ],
1602
- [
1603
- 2,
1604
- 768
1605
- ],
1606
- [
1607
- 2,
1608
- 768
1609
- ],
1610
- [
1611
- 4,
1612
- 768
1613
- ],
1614
- [
1615
- 1,
1616
- 768
1617
- ],
1618
- [
1619
- 2,
1620
- 768
1621
- ],
1622
- [
1623
- 2,
1624
- 768
1625
- ],
1626
- [
1627
- 2,
1628
- 768
1629
- ],
1630
- [
1631
- 3,
1632
- 768
1633
- ],
1634
- [
1635
- 3,
1636
- 768
1637
- ],
1638
- [
1639
- 2,
1640
- 768
1641
- ],
1642
- [
1643
- 2,
1644
- 768
1645
- ],
1646
- [
1647
- 1,
1648
- 768
1649
- ],
1650
- [
1651
- 13,
1652
- 768
1653
- ],
1654
- [
1655
- 3,
1656
- 768
1657
- ],
1658
- [
1659
- 5,
1660
- 768
1661
- ],
1662
- [
1663
- 11,
1664
- 768
1665
- ],
1666
- [
1667
- 3,
1668
- 768
1669
- ],
1670
- [
1671
- 37,
1672
- 768
1673
- ],
1674
- [
1675
- 2,
1676
- 768
1677
- ],
1678
- [
1679
- 49,
1680
- 768
1681
- ],
1682
- [
1683
- 12,
1684
- 768
1685
- ],
1686
- [
1687
- 40,
1688
- 768
1689
- ],
1690
- [
1691
- 10,
1692
- 768
1693
- ],
1694
- [
1695
- 4,
1696
- 768
1697
- ],
1698
- [
1699
- 1,
1700
- 768
1701
- ],
1702
- [
1703
- 2,
1704
- 768
1705
- ],
1706
- [
1707
- 2,
1708
- 768
1709
- ],
1710
- [
1711
- 1,
1712
- 768
1713
- ],
1714
- [
1715
- 5,
1716
- 768
1717
- ],
1718
- [
1719
- 2,
1720
- 768
1721
- ],
1722
- [
1723
- 2,
1724
- 768
1725
- ],
1726
- [
1727
- 1,
1728
- 768
1729
- ],
1730
- [
1731
- 3,
1732
- 768
1733
- ],
1734
- [
1735
- 2,
1736
- 768
1737
- ],
1738
- [
1739
- 3,
1740
- 768
1741
- ],
1742
- [
1743
- 2,
1744
- 768
1745
- ],
1746
- [
1747
- 2,
1748
- 768
1749
- ],
1750
- [
1751
- 2,
1752
- 768
1753
- ],
1754
- [
1755
- 2,
1756
- 768
1757
- ],
1758
- [
1759
- 2,
1760
- 768
1761
- ],
1762
- [
1763
- 3,
1764
- 768
1765
- ],
1766
- [
1767
- 2,
1768
- 768
1769
- ],
1770
- [
1771
- 2,
1772
- 768
1773
- ],
1774
- [
1775
- 12,
1776
- 768
1777
- ],
1778
- [
1779
- 3,
1780
- 768
1781
- ],
1782
- [
1783
- 3,
1784
- 768
1785
- ],
1786
- [
1787
- 2,
1788
- 768
1789
- ],
1790
- [
1791
- 19,
1792
- 768
1793
- ]
1794
  ],
1795
  "hidden_act": "gelu",
1796
  "hidden_dropout_prob": 0.1,
1797
  "hidden_size": 768,
 
 
 
 
 
1798
  "initializer_range": 0.02,
1799
  "intermediate_size": 3072,
 
 
 
 
 
1800
  "layer_norm_eps": 1e-07,
1801
  "max_position_embeddings": 512,
1802
  "max_relative_positions": -1,
@@ -1817,47 +454,39 @@
1817
  "relative_attention": true,
1818
  "share_att_key": true,
1819
  "tasks": [
1820
- "babi_nli/counting",
1821
- "babi_nli/indefinite-knowledge",
1822
- "babi_nli/simple-negation",
1823
- "babi_nli/three-arg-relations",
1824
- "babi_nli/basic-induction",
1825
- "babi_nli/time-reasoning",
1826
- "babi_nli/compound-coreference",
1827
- "babi_nli/path-finding",
1828
- "babi_nli/positional-reasoning",
1829
- "babi_nli/conjunction",
1830
- "babi_nli/size-reasoning",
1831
- "babi_nli/yes-no-questions",
1832
- "babi_nli/basic-coreference",
1833
- "babi_nli/two-supporting-facts",
1834
- "babi_nli/lists-sets",
1835
- "babi_nli/two-arg-relations",
1836
- "babi_nli/three-supporting-facts",
1837
- "babi_nli/basic-deduction",
1838
- "babi_nli/single-supporting-fact",
1839
  "anli/a1",
1840
  "anli/a2",
1841
  "anli/a3",
1842
  "sick/label",
1843
  "sick/relatedness",
1844
  "sick/entailment_AB",
1845
- "sick/entailment_BA",
1846
  "snli",
1847
  "scitail/snli_format",
1848
  "hans",
1849
  "WANLI",
1850
- "recast/recast_kg_relations",
1851
- "recast/recast_puns",
1852
  "recast/recast_factuality",
1853
- "recast/recast_megaveridicality",
1854
  "recast/recast_verbcorner",
1855
- "recast/recast_verbnet",
 
1856
  "recast/recast_ner",
1857
  "recast/recast_sentiment",
 
 
1858
  "probability_words_nli/usnli",
1859
- "probability_words_nli/reasoning_1hop",
1860
  "probability_words_nli/reasoning_2hop",
 
1861
  "nan-nli/joey234--nan-nli",
1862
  "nli_fever",
1863
  "breaking_nli",
@@ -1882,35 +511,13 @@
1882
  "robust_nli/ST_LM",
1883
  "robust_nli_is_sd",
1884
  "robust_nli_li_ts",
1885
- "gen_debiased_nli/snli_seq_z",
1886
- "gen_debiased_nli/snli_z_aug",
1887
- "gen_debiased_nli/snli_par_z",
1888
- "gen_debiased_nli/mnli_par_z",
1889
- "gen_debiased_nli/mnli_z_aug",
1890
- "gen_debiased_nli/mnli_seq_z",
1891
  "add_one_rte",
1892
- "imppres/presupposition_cleft_uniqueness/presupposition",
1893
- "imppres/presupposition_possessed_definites_uniqueness/presupposition",
1894
- "imppres/presupposition_possessed_definites_existence/presupposition",
1895
- "imppres/presupposition_only_presupposition/presupposition",
1896
- "imppres/presupposition_all_n_presupposition/presupposition",
1897
- "imppres/presupposition_both_presupposition/presupposition",
1898
- "imppres/presupposition_change_of_state/presupposition",
1899
- "imppres/presupposition_cleft_existence/presupposition",
1900
- "imppres/presupposition_question_presupposition/presupposition",
1901
- "imppres/implicature_modals/prag",
1902
- "imppres/implicature_numerals_10_100/prag",
1903
- "imppres/implicature_numerals_2_3/prag",
1904
- "imppres/implicature_gradable_adjective/prag",
1905
- "imppres/implicature_quantifiers/prag",
1906
- "imppres/implicature_gradable_verb/prag",
1907
- "imppres/implicature_connectives/prag",
1908
- "imppres/implicature_gradable_adjective/log",
1909
- "imppres/implicature_gradable_verb/log",
1910
  "imppres/implicature_numerals_2_3/log",
1911
  "imppres/implicature_numerals_10_100/log",
1912
  "imppres/implicature_modals/log",
1913
- "imppres/implicature_quantifiers/log",
 
1914
  "imppres/implicature_connectives/log",
1915
  "glue_diagnostics/diagnostics",
1916
  "hlgd",
@@ -1925,108 +532,106 @@
1925
  "model-written-evals",
1926
  "truthful_qa/multiple_choice",
1927
  "fig-qa",
1928
- "bigbench/fantasy_reasoning",
 
 
 
 
 
 
 
 
 
 
 
 
1929
  "bigbench/nonsense_words_grammar",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1930
  "bigbench/analytic_entailment",
1931
- "bigbench/logic_grid_puzzle",
1932
- "bigbench/geometric_shapes",
1933
- "bigbench/key_value_maps",
1934
- "bigbench/analogical_similarity",
1935
- "bigbench/metaphor_understanding",
 
 
 
 
 
 
1936
  "bigbench/metaphor_boolean",
1937
- "bigbench/ruin_names",
1938
- "bigbench/cs_algorithms",
1939
- "bigbench/physical_intuition",
1940
- "bigbench/mnist_ascii",
1941
- "bigbench/moral_permissibility",
 
 
 
 
 
 
 
1942
  "bigbench/emoji_movie",
1943
- "bigbench/snarks",
1944
- "bigbench/timedial",
1945
  "bigbench/dark_humor_detection",
1946
- "bigbench/gre_reading_comprehension",
1947
- "bigbench/empirical_judgments",
1948
- "bigbench/causal_judgment",
1949
- "bigbench/fact_checker",
1950
  "bigbench/logical_fallacy_detection",
1951
- "bigbench/identify_math_theorems",
1952
- "bigbench/dyck_languages",
1953
- "bigbench/winowhy",
1954
- "bigbench/logical_sequence",
1955
  "bigbench/strategyqa",
1956
- "bigbench/unit_interpretation",
1957
- "bigbench/authorship_verification",
1958
- "bigbench/undo_permutation",
1959
  "bigbench/epistemic_reasoning",
1960
- "bigbench/human_organs_senses",
1961
- "bigbench/misconceptions",
1962
- "bigbench/international_phonetic_alphabet_nli",
1963
- "bigbench/identify_odd_metaphor",
1964
- "bigbench/mathematical_induction",
1965
- "bigbench/odd_one_out",
1966
- "bigbench/reasoning_about_colored_objects",
1967
- "bigbench/strange_stories",
1968
- "bigbench/evaluating_information_essentiality",
1969
- "bigbench/figure_of_speech_detection",
1970
- "bigbench/english_proverbs",
1971
- "bigbench/general_knowledge",
1972
- "bigbench/tracking_shuffled_objects",
1973
- "bigbench/physics",
1974
  "bigbench/anachronisms",
 
 
 
 
1975
  "bigbench/simple_ethical_questions",
1976
- "bigbench/logical_args",
1977
- "bigbench/suicide_risk",
1978
- "bigbench/sentence_ambiguity",
1979
- "bigbench/temporal_sequences",
1980
- "bigbench/penguins_in_a_table",
1981
- "bigbench/sports_understanding",
1982
  "bigbench/hyperbaton",
1983
- "bigbench/code_line_description",
 
 
 
 
 
 
 
1984
  "bigbench/question_selection",
1985
- "bigbench/disambiguation_qa",
1986
- "bigbench/date_understanding",
1987
- "bigbench/play_dialog_same_or_different",
1988
- "bigbench/salient_translation_error_detection",
1989
- "bigbench/irony_identification",
1990
- "bigbench/emojis_emotion_prediction",
1991
- "bigbench/hindu_knowledge",
1992
- "bigbench/conceptual_combinations",
1993
  "bigbench/implicatures",
 
 
 
1994
  "bigbench/movie_dialog_same_or_different",
1995
- "bigbench/social_support",
1996
- "bigbench/presuppositions_as_nli",
1997
- "bigbench/vitaminc_fact_verification",
1998
- "bigbench/hhh_alignment",
1999
- "bigbench/implicit_relations",
2000
  "bigbench/bbq_lite_json",
2001
- "bigbench/phrase_relatedness",
2002
- "bigbench/logical_deduction",
2003
- "bigbench/discourse_marker_prediction",
2004
- "bigbench/movie_recommendation",
2005
- "bigbench/real_or_fake_text",
2006
- "bigbench/formal_fallacies_syllogisms_negation",
2007
- "bigbench/crass_ai",
2008
- "blimp/inchoative",
2009
- "blimp/principle_A_c_command",
2010
- "blimp/matrix_question_npi_licensor_present",
2011
- "blimp/wh_questions_subject_gap_long_distance",
2012
- "blimp/sentential_subject_island",
2013
- "blimp/existential_there_quantifiers_2",
2014
- "blimp/sentential_negation_npi_scope",
2015
- "blimp/complex_NP_island",
2016
- "blimp/principle_A_reconstruction",
2017
- "blimp/animate_subject_passive",
2018
- "blimp/tough_vs_raising_1",
2019
- "blimp/wh_vs_that_with_gap",
2020
- "blimp/principle_A_domain_2",
2021
- "blimp/npi_present_1",
2022
- "blimp/wh_vs_that_with_gap_long_distance",
2023
- "blimp/superlative_quantifiers_1",
2024
- "blimp/npi_present_2",
2025
- "blimp/wh_questions_object_gap",
2026
- "blimp/coordinate_structure_constraint_complex_left_branch",
2027
- "blimp/coordinate_structure_constraint_object_extraction",
2028
- "blimp/left_branch_island_echo_question",
2029
- "blimp/drop_argument",
2030
  "cos_e/v1.0",
2031
  "cosmos_qa",
2032
  "dream",
@@ -2042,72 +647,22 @@
2042
  "piqa",
2043
  "hellaswag",
2044
  "super_glue/copa",
 
 
2045
  "art",
2046
- "hendrycks_test/moral_disputes",
2047
- "hendrycks_test/moral_scenarios",
2048
- "hendrycks_test/nutrition",
2049
- "hendrycks_test/philosophy",
2050
- "hendrycks_test/prehistory",
2051
- "hendrycks_test/professional_accounting",
2052
- "hendrycks_test/professional_law",
2053
- "hendrycks_test/world_religions",
2054
- "hendrycks_test/professional_psychology",
2055
- "hendrycks_test/public_relations",
2056
- "hendrycks_test/security_studies",
2057
- "hendrycks_test/sociology",
2058
- "hendrycks_test/us_foreign_policy",
2059
- "hendrycks_test/virology",
2060
- "hendrycks_test/miscellaneous",
2061
- "hendrycks_test/professional_medicine",
2062
- "hendrycks_test/medical_genetics",
2063
- "hendrycks_test/college_mathematics",
2064
- "hendrycks_test/management",
2065
- "hendrycks_test/high_school_computer_science",
2066
- "hendrycks_test/astronomy",
2067
- "hendrycks_test/high_school_chemistry",
2068
- "hendrycks_test/high_school_biology",
2069
- "hendrycks_test/global_facts",
2070
- "hendrycks_test/formal_logic",
2071
- "hendrycks_test/elementary_mathematics",
2072
- "hendrycks_test/high_school_european_history",
2073
- "hendrycks_test/electrical_engineering",
2074
- "hendrycks_test/conceptual_physics",
2075
- "hendrycks_test/computer_security",
2076
- "hendrycks_test/college_physics",
2077
- "hendrycks_test/college_medicine",
2078
- "hendrycks_test/college_computer_science",
2079
- "hendrycks_test/college_chemistry",
2080
- "hendrycks_test/college_biology",
2081
- "hendrycks_test/econometrics",
2082
- "hendrycks_test/clinical_knowledge",
2083
- "hendrycks_test/anatomy",
2084
- "hendrycks_test/marketing",
2085
- "hendrycks_test/machine_learning",
2086
- "hendrycks_test/logical_fallacies",
2087
- "hendrycks_test/jurisprudence",
2088
- "hendrycks_test/international_law",
2089
- "hendrycks_test/human_sexuality",
2090
- "hendrycks_test/human_aging",
2091
- "hendrycks_test/high_school_world_history",
2092
- "hendrycks_test/abstract_algebra",
2093
- "hendrycks_test/high_school_us_history",
2094
- "hendrycks_test/high_school_psychology",
2095
- "hendrycks_test/high_school_physics",
2096
- "hendrycks_test/high_school_microeconomics",
2097
- "hendrycks_test/high_school_mathematics",
2098
- "hendrycks_test/high_school_macroeconomics",
2099
- "hendrycks_test/high_school_government_and_politics",
2100
- "hendrycks_test/high_school_geography",
2101
- "hendrycks_test/high_school_statistics",
2102
- "hendrycks_test/business_ethics",
2103
  "winogrande/winogrande_xl",
2104
  "codah/codah",
2105
- "ai2_arc/ARC-Challenge/challenge",
2106
  "ai2_arc/ARC-Easy/challenge",
 
2107
  "definite_pronoun_resolution",
2108
  "swag",
2109
  "math_qa",
 
 
2110
  "utilitarianism",
 
 
 
2111
  "TuringBench",
2112
  "trec",
2113
  "vitaminc/tals--vitaminc",
@@ -2115,57 +670,52 @@
2115
  "rumoureval_2019/RumourEval2019",
2116
  "ethos/binary",
2117
  "ethos/multilabel",
2118
- "glue/cola",
2119
- "glue/sst2",
2120
- "glue/mrpc",
2121
- "glue/qqp",
2122
- "glue/stsb",
2123
- "glue/mnli",
2124
- "glue/qnli",
2125
- "glue/rte",
2126
- "glue/wnli",
2127
- "super_glue/boolq",
2128
- "super_glue/cb",
2129
- "super_glue/multirc",
2130
- "super_glue/wic",
2131
- "super_glue/axg",
2132
- "tweet_eval/stance_feminist",
2133
  "tweet_eval/stance_atheism",
 
2134
  "tweet_eval/stance_hillary",
 
2135
  "tweet_eval/stance_abortion",
2136
- "tweet_eval/sentiment",
2137
- "tweet_eval/offensive",
2138
- "tweet_eval/stance_climate",
2139
- "tweet_eval/irony",
2140
- "tweet_eval/emotion",
2141
- "tweet_eval/emoji",
2142
- "tweet_eval/hate",
2143
  "discovery/discovery",
2144
- "pragmeval/switchboard",
2145
  "pragmeval/squinky-informativeness",
2146
- "pragmeval/emobank-arousal",
2147
- "pragmeval/emobank-dominance",
2148
- "pragmeval/emobank-valence",
2149
- "pragmeval/mrda",
2150
  "pragmeval/verifiability",
2151
- "pragmeval/squinky-implicature",
2152
  "pragmeval/squinky-formality",
2153
- "pragmeval/gum",
2154
- "pragmeval/emergent",
 
 
 
 
2155
  "pragmeval/persuasiveness-premisetype",
 
2156
  "pragmeval/pdtb",
2157
- "pragmeval/persuasiveness-eloquence",
2158
- "pragmeval/persuasiveness-specificity",
2159
  "pragmeval/persuasiveness-strength",
2160
  "pragmeval/sarcasm",
2161
- "pragmeval/stac",
 
2162
  "pragmeval/persuasiveness-claimtype",
2163
- "pragmeval/persuasiveness-relevance",
 
 
 
 
 
 
 
 
 
2164
  "lex_glue/eurlex",
2165
  "lex_glue/scotus",
2166
  "lex_glue/ledgar",
2167
  "lex_glue/unfair_tos",
2168
  "lex_glue/case_hold",
 
2169
  "imdb",
2170
  "rotten_tomatoes",
2171
  "ag_news",
@@ -2182,34 +732,36 @@
2182
  "snips_built_in_intents",
2183
  "banking77",
2184
  "hate_speech_offensive",
2185
- "hyperpartisan_news_detection/byarticle",
2186
- "hyperpartisan_news_detection/bypublisher",
 
 
 
2187
  "go_emotions/simplified",
2188
  "scicite",
2189
  "liar",
2190
- "lexical_relation_classification/ROOT09",
2191
- "lexical_relation_classification/EVALution",
2192
  "lexical_relation_classification/CogALexV",
2193
- "lexical_relation_classification/BLESS",
2194
  "lexical_relation_classification/K&H+N",
2195
- "linguisticprobing/coordination_inversion",
2196
- "linguisticprobing/odd_man_out",
2197
- "linguisticprobing/word_content",
2198
- "linguisticprobing/obj_number",
2199
- "linguisticprobing/past_present",
2200
- "linguisticprobing/tree_depth",
2201
- "linguisticprobing/sentence_length",
2202
  "linguisticprobing/top_constituents",
2203
  "linguisticprobing/bigram_shift",
 
 
 
 
 
2204
  "linguisticprobing/subj_number",
 
 
 
 
2205
  "crowdflower/sentiment_nuclear_power",
 
 
2206
  "crowdflower/tweet_global_warming",
2207
  "crowdflower/airline-sentiment",
2208
- "crowdflower/economic-news",
2209
- "crowdflower/political-media-audience",
2210
- "crowdflower/political-media-bias",
2211
- "crowdflower/political-media-message",
2212
- "crowdflower/text_emotion",
2213
  "crowdflower/corporate-messaging",
2214
  "ethics/commonsense",
2215
  "ethics/deontology",
@@ -2218,18 +770,11 @@
2218
  "emo/emo2019",
2219
  "google_wellformed_query",
2220
  "tweets_hate_speech_detection",
2221
- "adv_glue/adv_sst2",
2222
- "adv_glue/adv_qqp",
2223
- "adv_glue/adv_mnli",
2224
- "adv_glue/adv_mnli_mismatched",
2225
- "adv_glue/adv_qnli",
2226
- "adv_glue/adv_rte",
2227
  "has_part",
2228
  "wnut_17/wnut_17",
2229
  "ncbi_disease/ncbi_disease",
2230
  "acronym_identification",
2231
  "jnlpba/jnlpba",
2232
- "species_800/species_800",
2233
  "ontonotes_english/SpeedOfMagic--ontonotes_english",
2234
  "blog_authorship_corpus/gender",
2235
  "blog_authorship_corpus/age",
@@ -2242,9 +787,6 @@
2242
  "ade_corpus_v2/Ade_corpus_v2_classification",
2243
  "discosense",
2244
  "circa",
2245
- "code_x_glue_cc_defect_detection",
2246
- "code_x_glue_cc_clone_detection_big_clone_bench",
2247
- "code_x_glue_cc_code_refinement/medium",
2248
  "EffectiveFeedbackStudentWriting",
2249
  "promptSentiment",
2250
  "promptNLI",
@@ -2260,10 +802,74 @@
2260
  "dynasent/dynabench.dynasent.r1.all/r1",
2261
  "dynasent/dynabench.dynasent.r2.all/r2",
2262
  "Sarcasm_News_Headline",
2263
- "sem_eval_2010_task_8"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2264
  ],
2265
  "torch_dtype": "float32",
2266
- "transformers_version": "4.25.1",
2267
  "type_vocab_size": 0,
2268
  "vocab_size": 128100
2269
  }
 
1
  {
2
+ "_name_or_path": "microsoft/deberta-v3-base",
3
  "architectures": [
4
  "DebertaV2ForSequenceClassification"
5
  ],
 
 
 
 
 
 
 
 
 
 
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifiers_size": [
8
+ 3,
9
+ 2,
10
+ 2,
11
+ 2,
12
+ 2,
13
+ 2,
14
+ 1,
15
+ 2,
16
+ 3,
17
+ 2,
18
+ 2,
19
+ 2,
20
+ 3,
21
+ 3,
22
+ 3,
23
+ 3,
24
+ 1,
25
+ 3,
26
+ 3,
27
+ 2,
28
+ 2,
29
+ 3,
30
+ 2,
31
+ 2,
32
+ 2,
33
+ 6,
34
+ 2,
35
+ 2,
36
+ 2,
37
+ 2,
38
+ 2,
39
+ 2,
40
+ 2,
41
+ 3,
42
+ 3,
43
+ 3,
44
+ 3,
45
+ 3,
46
+ 3,
47
+ 3,
48
+ 2,
49
+ 3,
50
+ 2,
51
+ 2,
52
+ 2,
53
+ 3,
54
+ 3,
55
+ 3,
56
+ 3,
57
+ 3,
58
+ 3,
59
+ 3,
60
+ 3,
61
+ 3,
62
+ 3,
63
+ 2,
64
+ 2,
65
+ 2,
66
+ 3,
67
+ 3,
68
+ 3,
69
+ 3,
70
+ 3,
71
+ 3,
72
+ 3,
73
+ 3,
74
+ 2,
75
+ 2,
76
+ 2,
77
+ 2,
78
+ 2,
79
+ 47,
80
+ 23,
81
+ 9,
82
+ 1,
83
+ 1,
84
+ 1,
85
+ 1,
86
+ 1,
87
+ 1,
88
+ 1,
89
+ 1,
90
+ 1,
91
+ 1,
92
+ 1,
93
+ 1,
94
+ 1,
95
+ 1,
96
+ 1,
97
+ 1,
98
+ 1,
99
+ 1,
100
+ 1,
101
+ 1,
102
+ 1,
103
+ 1,
104
+ 1,
105
+ 1,
106
+ 1,
107
+ 1,
108
+ 1,
109
+ 1,
110
+ 1,
111
+ 1,
112
+ 1,
113
+ 1,
114
+ 1,
115
+ 1,
116
+ 1,
117
+ 1,
118
+ 1,
119
+ 1,
120
+ 1,
121
+ 1,
122
+ 1,
123
+ 1,
124
+ 1,
125
+ 1,
126
+ 1,
127
+ 1,
128
+ 1,
129
+ 1,
130
+ 1,
131
+ 1,
132
+ 1,
133
+ 1,
134
+ 1,
135
+ 1,
136
+ 1,
137
+ 1,
138
+ 1,
139
+ 1,
140
+ 1,
141
+ 1,
142
+ 1,
143
+ 1,
144
+ 1,
145
+ 1,
146
+ 1,
147
+ 1,
148
+ 1,
149
+ 1,
150
+ 1,
151
+ 1,
152
+ 1,
153
+ 1,
154
+ 1,
155
+ 1,
156
+ 1,
157
+ 1,
158
+ 1,
159
+ 1,
160
+ 1,
161
+ 1,
162
+ 1,
163
+ 1,
164
+ 1,
165
+ 1,
166
+ 1,
167
+ 1,
168
+ 1,
169
+ 1,
170
+ 1,
171
+ 1,
172
+ 1,
173
+ 1,
174
+ 1,
175
+ 1,
176
+ 1,
177
+ 1,
178
+ 1,
179
+ 1,
180
+ 1,
181
+ 1,
182
+ 1,
183
+ 1,
184
+ 1,
185
+ 1,
186
+ 1,
187
+ 1,
188
+ 1,
189
+ 1,
190
+ 1,
191
+ 1,
192
+ 1,
193
+ 1,
194
+ 1,
195
+ 1,
196
+ 1,
197
+ 1,
198
+ 1,
199
+ 1,
200
+ 1,
201
+ 1,
202
+ 1,
203
+ 1,
204
+ 1,
205
+ 1,
206
+ 1,
207
+ 1,
208
+ 1,
209
+ 1,
210
+ 1,
211
+ 2,
212
+ 2,
213
+ 2,
214
+ 2,
215
+ 2,
216
+ 2,
217
+ 20,
218
+ 50,
219
+ 3,
220
+ 3,
221
+ 4,
222
+ 2,
223
+ 8,
224
+ 20,
225
+ 4,
226
+ 2,
227
+ 2,
228
+ 3,
229
+ 2,
230
+ 3,
231
+ 3,
232
+ 3,
233
+ 3,
234
+ 3,
235
+ 174,
236
+ 2,
237
+ 3,
238
+ 2,
239
+ 2,
240
+ 2,
241
+ 2,
242
+ 41,
243
+ 51,
244
+ 2,
245
+ 8,
246
+ 2,
247
+ 16,
248
+ 18,
249
+ 2,
250
+ 2,
251
+ 17,
252
+ 2,
253
+ 3,
254
+ 2,
255
+ 3,
256
+ 12,
257
+ 42,
258
+ 3,
259
+ 7,
260
+ 11,
261
+ 7,
262
+ 4,
263
+ 3,
264
+ 100,
265
+ 13,
266
+ 100,
267
+ 8,
268
+ 1,
269
+ 20,
270
+ 2,
271
+ 2,
272
+ 4,
273
+ 5,
274
+ 3,
275
+ 4,
276
+ 14,
277
+ 2,
278
+ 6,
279
+ 4,
280
+ 2,
281
+ 1,
282
+ 3,
283
+ 10,
284
+ 77,
285
+ 3,
286
+ 10,
287
+ 4,
288
+ 2,
289
+ 7,
290
+ 6,
291
+ 28,
292
+ 3,
293
+ 6,
294
+ 5,
295
+ 7,
296
+ 4,
297
+ 3,
298
+ 6,
299
+ 20,
300
+ 2,
301
+ 2,
302
+ 2,
303
+ 6,
304
+ 2,
305
+ 2,
306
+ 2,
307
+ 7,
308
+ 3,
309
+ 13,
310
+ 9,
311
+ 4,
312
+ 2,
313
+ 2,
314
+ 2,
315
+ 3,
316
+ 4,
317
+ 2,
318
+ 2,
319
+ 2,
320
+ 2,
321
+ 4,
322
+ 1,
323
+ 2,
324
+ 1,
325
+ 13,
326
+ 3,
327
+ 5,
328
+ 11,
329
+ 37,
330
+ 2,
331
+ 49,
332
+ 12,
333
+ 40,
334
+ 10,
335
+ 4,
336
+ 1,
337
+ 2,
338
+ 2,
339
+ 1,
340
+ 5,
341
+ 3,
342
+ 2,
343
+ 3,
344
+ 2,
345
+ 2,
346
+ 2,
347
+ 2,
348
+ 2,
349
+ 3,
350
+ 2,
351
+ 2,
352
+ 12,
353
+ 3,
354
+ 3,
355
+ 2,
356
+ 19,
357
+ 3,
358
+ 1,
359
+ 1,
360
+ 2,
361
+ 2,
362
+ 2,
363
+ 2,
364
+ 2,
365
+ 1,
366
+ 2,
367
+ 2,
368
+ 1,
369
+ 1,
370
+ 2,
371
+ 3,
372
+ 2,
373
+ 1,
374
+ 4,
375
+ 4,
376
+ 1,
377
+ 1,
378
+ 1,
379
+ 2,
380
+ 3,
381
+ 2,
382
+ 3,
383
+ 1,
384
+ 1,
385
+ 2,
386
+ 1,
387
+ 3,
388
+ 2,
389
+ 2,
390
+ 2,
391
+ 2,
392
+ 3,
393
+ 2,
394
+ 2,
395
+ 2,
396
+ 1,
397
+ 3,
398
+ 2,
399
+ 2,
400
+ 1,
401
+ 1,
402
+ 1,
403
+ 1,
404
+ 2,
405
+ 1,
406
+ 1,
407
+ 1,
408
+ 1,
409
+ 4,
410
+ 1,
411
+ 1,
412
+ 1,
413
+ 3,
414
+ 1,
415
+ 2,
416
+ 3,
417
+ 3,
418
+ 3,
419
+ 1,
420
+ 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
421
  ],
422
  "hidden_act": "gelu",
423
  "hidden_dropout_prob": 0.1,
424
  "hidden_size": 768,
425
+ "id2label": {
426
+ "0": "entailment",
427
+ "1": "neutral",
428
+ "2": "contradiction"
429
+ },
430
  "initializer_range": 0.02,
431
  "intermediate_size": 3072,
432
+ "label2id": {
433
+ "contradiction": 2,
434
+ "entailment": 0,
435
+ "neutral": 1
436
+ },
437
  "layer_norm_eps": 1e-07,
438
  "max_position_embeddings": 512,
439
  "max_relative_positions": -1,
 
454
  "relative_attention": true,
455
  "share_att_key": true,
456
  "tasks": [
457
+ "glue/mnli",
458
+ "glue/qnli",
459
+ "glue/rte",
460
+ "glue/wnli",
461
+ "glue/mrpc",
462
+ "glue/qqp",
463
+ "glue/stsb",
464
+ "super_glue/boolq",
465
+ "super_glue/cb",
466
+ "super_glue/multirc",
467
+ "super_glue/wic",
468
+ "super_glue/axg",
 
 
 
 
 
 
 
469
  "anli/a1",
470
  "anli/a2",
471
  "anli/a3",
472
  "sick/label",
473
  "sick/relatedness",
474
  "sick/entailment_AB",
 
475
  "snli",
476
  "scitail/snli_format",
477
  "hans",
478
  "WANLI",
 
 
479
  "recast/recast_factuality",
 
480
  "recast/recast_verbcorner",
481
+ "recast/recast_puns",
482
+ "recast/recast_kg_relations",
483
  "recast/recast_ner",
484
  "recast/recast_sentiment",
485
+ "recast/recast_megaveridicality",
486
+ "recast/recast_verbnet",
487
  "probability_words_nli/usnli",
 
488
  "probability_words_nli/reasoning_2hop",
489
+ "probability_words_nli/reasoning_1hop",
490
  "nan-nli/joey234--nan-nli",
491
  "nli_fever",
492
  "breaking_nli",
 
511
  "robust_nli/ST_LM",
512
  "robust_nli_is_sd",
513
  "robust_nli_li_ts",
 
 
 
 
 
 
514
  "add_one_rte",
515
+ "imppres/implicature_quantifiers/log",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
516
  "imppres/implicature_numerals_2_3/log",
517
  "imppres/implicature_numerals_10_100/log",
518
  "imppres/implicature_modals/log",
519
+ "imppres/implicature_gradable_verb/log",
520
+ "imppres/implicature_gradable_adjective/log",
521
  "imppres/implicature_connectives/log",
522
  "glue_diagnostics/diagnostics",
523
  "hlgd",
 
532
  "model-written-evals",
533
  "truthful_qa/multiple_choice",
534
  "fig-qa",
535
+ "bigbench/similarities_abstraction",
536
+ "bigbench/tracking_shuffled_objects",
537
+ "bigbench/dyck_languages",
538
+ "bigbench/cause_and_effect",
539
+ "bigbench/contextual_parametric_knowledge_conflicts",
540
+ "bigbench/logical_deduction",
541
+ "bigbench/movie_recommendation",
542
+ "bigbench/hhh_alignment",
543
+ "bigbench/checkmate_in_one",
544
+ "bigbench/discourse_marker_prediction",
545
+ "bigbench/identify_odd_metaphor",
546
+ "bigbench/entailed_polarity",
547
+ "bigbench/novel_concepts",
548
  "bigbench/nonsense_words_grammar",
549
+ "bigbench/goal_step_wikihow",
550
+ "bigbench/sports_understanding",
551
+ "bigbench/moral_permissibility",
552
+ "bigbench/code_line_description",
553
+ "bigbench/odd_one_out",
554
+ "bigbench/empirical_judgments",
555
+ "bigbench/gre_reading_comprehension",
556
+ "bigbench/reasoning_about_colored_objects",
557
+ "bigbench/identify_math_theorems",
558
+ "bigbench/abstract_narrative_understanding",
559
+ "bigbench/date_understanding",
560
+ "bigbench/play_dialog_same_or_different",
561
+ "bigbench/real_or_fake_text",
562
+ "bigbench/crass_ai",
563
+ "bigbench/international_phonetic_alphabet_nli",
564
+ "bigbench/presuppositions_as_nli",
565
+ "bigbench/ruin_names",
566
+ "bigbench/sentence_ambiguity",
567
+ "bigbench/intent_recognition",
568
  "bigbench/analytic_entailment",
569
+ "bigbench/conceptual_combinations",
570
+ "bigbench/implicit_relations",
571
+ "bigbench/emojis_emotion_prediction",
572
+ "bigbench/hindu_knowledge",
573
+ "bigbench/elementary_math_qa",
574
+ "bigbench/mathematical_induction",
575
+ "bigbench/irony_identification",
576
+ "bigbench/authorship_verification",
577
+ "bigbench/disambiguation_qa",
578
+ "bigbench/general_knowledge",
579
+ "bigbench/known_unknowns",
580
  "bigbench/metaphor_boolean",
581
+ "bigbench/logical_args",
582
+ "bigbench/metaphor_understanding",
583
+ "bigbench/arithmetic",
584
+ "bigbench/salient_translation_error_detection",
585
+ "bigbench/human_organs_senses",
586
+ "bigbench/formal_fallacies_syllogisms_negation",
587
+ "bigbench/fantasy_reasoning",
588
+ "bigbench/fact_checker",
589
+ "bigbench/unit_interpretation",
590
+ "bigbench/vitaminc_fact_verification",
591
+ "bigbench/cifar10_classification",
592
+ "bigbench/social_iqa",
593
  "bigbench/emoji_movie",
 
 
594
  "bigbench/dark_humor_detection",
595
+ "bigbench/understanding_fables",
596
+ "bigbench/symbol_interpretation",
597
+ "bigbench/social_support",
 
598
  "bigbench/logical_fallacy_detection",
599
+ "bigbench/causal_judgment",
 
 
 
600
  "bigbench/strategyqa",
 
 
 
601
  "bigbench/epistemic_reasoning",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
602
  "bigbench/anachronisms",
603
+ "bigbench/mnist_ascii",
604
+ "bigbench/crash_blossom",
605
+ "bigbench/english_proverbs",
606
+ "bigbench/cs_algorithms",
607
  "bigbench/simple_ethical_questions",
 
 
 
 
 
 
608
  "bigbench/hyperbaton",
609
+ "bigbench/key_value_maps",
610
+ "bigbench/physics",
611
+ "bigbench/figure_of_speech_detection",
612
+ "bigbench/undo_permutation",
613
+ "bigbench/timedial",
614
+ "bigbench/geometric_shapes",
615
+ "bigbench/analogical_similarity",
616
+ "bigbench/color",
617
  "bigbench/question_selection",
618
+ "bigbench/evaluating_information_essentiality",
619
+ "bigbench/phrase_relatedness",
620
+ "bigbench/logic_grid_puzzle",
621
+ "bigbench/misconceptions",
622
+ "bigbench/winowhy",
623
+ "bigbench/snarks",
624
+ "bigbench/temporal_sequences",
625
+ "bigbench/riddle_sense",
626
  "bigbench/implicatures",
627
+ "bigbench/logical_sequence",
628
+ "bigbench/suicide_risk",
629
+ "bigbench/physical_intuition",
630
  "bigbench/movie_dialog_same_or_different",
631
+ "bigbench/navigate",
632
+ "bigbench/penguins_in_a_table",
633
+ "bigbench/strange_stories",
 
 
634
  "bigbench/bbq_lite_json",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
635
  "cos_e/v1.0",
636
  "cosmos_qa",
637
  "dream",
 
647
  "piqa",
648
  "hellaswag",
649
  "super_glue/copa",
650
+ "balanced-copa",
651
+ "e-CARE",
652
  "art",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
653
  "winogrande/winogrande_xl",
654
  "codah/codah",
 
655
  "ai2_arc/ARC-Easy/challenge",
656
+ "ai2_arc/ARC-Challenge/challenge",
657
  "definite_pronoun_resolution",
658
  "swag",
659
  "math_qa",
660
+ "glue/cola",
661
+ "glue/sst2",
662
  "utilitarianism",
663
+ "amazon_counterfactual/en",
664
+ "insincere-questions",
665
+ "toxic_conversations",
666
  "TuringBench",
667
  "trec",
668
  "vitaminc/tals--vitaminc",
 
670
  "rumoureval_2019/RumourEval2019",
671
  "ethos/binary",
672
  "ethos/multilabel",
673
+ "tweet_eval/emoji",
674
+ "tweet_eval/emotion",
675
+ "tweet_eval/hate",
676
+ "tweet_eval/offensive",
677
+ "tweet_eval/sentiment",
678
+ "tweet_eval/irony",
 
 
 
 
 
 
 
 
 
679
  "tweet_eval/stance_atheism",
680
+ "tweet_eval/stance_climate",
681
  "tweet_eval/stance_hillary",
682
+ "tweet_eval/stance_feminist",
683
  "tweet_eval/stance_abortion",
 
 
 
 
 
 
 
684
  "discovery/discovery",
 
685
  "pragmeval/squinky-informativeness",
 
 
 
 
686
  "pragmeval/verifiability",
 
687
  "pragmeval/squinky-formality",
688
+ "pragmeval/squinky-implicature",
689
+ "pragmeval/emobank-dominance",
690
+ "pragmeval/emobank-arousal",
691
+ "pragmeval/switchboard",
692
+ "pragmeval/mrda",
693
+ "pragmeval/emobank-valence",
694
  "pragmeval/persuasiveness-premisetype",
695
+ "pragmeval/persuasiveness-relevance",
696
  "pragmeval/pdtb",
697
+ "pragmeval/stac",
 
698
  "pragmeval/persuasiveness-strength",
699
  "pragmeval/sarcasm",
700
+ "pragmeval/gum",
701
+ "pragmeval/persuasiveness-eloquence",
702
  "pragmeval/persuasiveness-claimtype",
703
+ "pragmeval/persuasiveness-specificity",
704
+ "pragmeval/emergent",
705
+ "silicone/maptask",
706
+ "silicone/oasis",
707
+ "silicone/meld_s",
708
+ "silicone/meld_e",
709
+ "silicone/iemocap",
710
+ "silicone/dyda_e",
711
+ "silicone/dyda_da",
712
+ "silicone/sem",
713
  "lex_glue/eurlex",
714
  "lex_glue/scotus",
715
  "lex_glue/ledgar",
716
  "lex_glue/unfair_tos",
717
  "lex_glue/case_hold",
718
+ "language-identification",
719
  "imdb",
720
  "rotten_tomatoes",
721
  "ag_news",
 
732
  "snips_built_in_intents",
733
  "banking77",
734
  "hate_speech_offensive",
735
+ "yahoo_answers_topics",
736
+ "stackoverflow-questions",
737
+ "hyperpartisan_news",
738
+ "sciie",
739
+ "citation_intent",
740
  "go_emotions/simplified",
741
  "scicite",
742
  "liar",
 
 
743
  "lexical_relation_classification/CogALexV",
744
+ "lexical_relation_classification/EVALution",
745
  "lexical_relation_classification/K&H+N",
746
+ "lexical_relation_classification/ROOT09",
747
+ "lexical_relation_classification/BLESS",
 
 
 
 
 
748
  "linguisticprobing/top_constituents",
749
  "linguisticprobing/bigram_shift",
750
+ "linguisticprobing/odd_man_out",
751
+ "linguisticprobing/coordination_inversion",
752
+ "linguisticprobing/sentence_length",
753
+ "linguisticprobing/past_present",
754
+ "linguisticprobing/obj_number",
755
  "linguisticprobing/subj_number",
756
+ "linguisticprobing/tree_depth",
757
+ "crowdflower/economic-news",
758
+ "crowdflower/text_emotion",
759
+ "crowdflower/political-media-message",
760
  "crowdflower/sentiment_nuclear_power",
761
+ "crowdflower/political-media-bias",
762
+ "crowdflower/political-media-audience",
763
  "crowdflower/tweet_global_warming",
764
  "crowdflower/airline-sentiment",
 
 
 
 
 
765
  "crowdflower/corporate-messaging",
766
  "ethics/commonsense",
767
  "ethics/deontology",
 
770
  "emo/emo2019",
771
  "google_wellformed_query",
772
  "tweets_hate_speech_detection",
 
 
 
 
 
 
773
  "has_part",
774
  "wnut_17/wnut_17",
775
  "ncbi_disease/ncbi_disease",
776
  "acronym_identification",
777
  "jnlpba/jnlpba",
 
778
  "ontonotes_english/SpeedOfMagic--ontonotes_english",
779
  "blog_authorship_corpus/gender",
780
  "blog_authorship_corpus/age",
 
787
  "ade_corpus_v2/Ade_corpus_v2_classification",
788
  "discosense",
789
  "circa",
 
 
 
790
  "EffectiveFeedbackStudentWriting",
791
  "promptSentiment",
792
  "promptNLI",
 
802
  "dynasent/dynabench.dynasent.r1.all/r1",
803
  "dynasent/dynabench.dynasent.r2.all/r2",
804
  "Sarcasm_News_Headline",
805
+ "sem_eval_2010_task_8",
806
+ "auditor_review/demo-org--auditor_review",
807
+ "medmcqa",
808
+ "aqua_rat/tokenized",
809
+ "Dynasent_Disagreement",
810
+ "Politeness_Disagreement",
811
+ "SBIC_Disagreement",
812
+ "SChem_Disagreement",
813
+ "Dilemmas_Disagreement",
814
+ "logiqa",
815
+ "wiki_qa",
816
+ "cycic_classification",
817
+ "cycic_multiplechoice",
818
+ "sts-companion",
819
+ "commonsense_qa_2.0",
820
+ "lingnli",
821
+ "monotonicity-entailment",
822
+ "arct",
823
+ "scinli",
824
+ "naturallogic",
825
+ "onestop_qa",
826
+ "moral_stories/full",
827
+ "prost",
828
+ "dynahate",
829
+ "syntactic-augmentation-nli",
830
+ "autotnli",
831
+ "CONDAQA",
832
+ "webgpt_comparisons",
833
+ "synthetic-instruct-gptj-pairwise",
834
+ "scruples",
835
+ "wouldyourather",
836
+ "attempto-nli",
837
+ "defeasible-nli/atomic",
838
+ "defeasible-nli/snli",
839
+ "nli-veridicality-transitivity",
840
+ "natural-language-satisfiability",
841
+ "lonli",
842
+ "dadc-limit-nli",
843
+ "FLUTE",
844
+ "strategy-qa",
845
+ "summarize_from_feedback/comparisons",
846
+ "folio",
847
+ "tomi-nli",
848
+ "avicenna",
849
+ "SHP",
850
+ "MedQA-USMLE-4-options-hf",
851
+ "wikimedqa/medwiki",
852
+ "cicero",
853
+ "CREAK",
854
+ "mutual",
855
+ "NeQA",
856
+ "quote-repetition",
857
+ "redefine-math",
858
+ "puzzte",
859
+ "implicatures",
860
+ "race/high",
861
+ "race/middle",
862
+ "spartqa-yn",
863
+ "spartqa-mchoice",
864
+ "babi_nli",
865
+ "gen_debiased_nli",
866
+ "imppres/presupposition",
867
+ "/prag",
868
+ "blimp-2",
869
+ "mmlu-4"
870
  ],
871
  "torch_dtype": "float32",
872
+ "transformers_version": "4.26.1",
873
  "type_vocab_size": 0,
874
  "vocab_size": 128100
875
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c9c0c119413c34bf7b6af3a5b22bd4916e9a23d72933e21ea97c44f83adb5b78
3
- size 737772498
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75ffa91a8efd68439bb0a54efcc2d38a7f121085f1782587645495b0016a464d
3
+ size 737771833