Aswanth-Azma commited on
Commit
d9ea848
1 Parent(s): 0c85c74

updated readme

Browse files
Files changed (1) hide show
  1. README.md +0 -285
README.md CHANGED
@@ -1,288 +1,3 @@
1
- ---
2
- license: apache-2.0
3
- language: en
4
- tags:
5
- - deberta-v3-base
6
- - deberta-v3
7
- - deberta
8
- - text-classification
9
- - nli
10
- - natural-language-inference
11
- - multitask
12
- - multi-task
13
- - pipeline
14
- - extreme-multi-task
15
- - extreme-mtl
16
- - tasksource
17
- - zero-shot
18
- - rlhf
19
- model-index:
20
- - name: deberta-v3-base-tasksource-nli
21
- results:
22
- - task:
23
- type: text-classification
24
- name: Text Classification
25
- dataset:
26
- name: glue
27
- type: glue
28
- config: rte
29
- split: validation
30
- metrics:
31
- - type: accuracy
32
- value: 0.89
33
- - task:
34
- type: natural-language-inference
35
- name: Natural Language Inference
36
- dataset:
37
- name: anli-r3
38
- type: anli
39
- config: plain_text
40
- split: validation
41
- metrics:
42
- - type: accuracy
43
- value: 0.52
44
- name: Accuracy
45
- datasets:
46
- - glue
47
- - super_glue
48
- - anli
49
- - tasksource/babi_nli
50
- - sick
51
- - snli
52
- - scitail
53
- - OpenAssistant/oasst1
54
- - universal_dependencies
55
- - hans
56
- - qbao775/PARARULE-Plus
57
- - alisawuffles/WANLI
58
- - metaeval/recast
59
- - sileod/probability_words_nli
60
- - joey234/nan-nli
61
- - pietrolesci/nli_fever
62
- - pietrolesci/breaking_nli
63
- - pietrolesci/conj_nli
64
- - pietrolesci/fracas
65
- - pietrolesci/dialogue_nli
66
- - pietrolesci/mpe
67
- - pietrolesci/dnc
68
- - pietrolesci/gpt3_nli
69
- - pietrolesci/recast_white
70
- - pietrolesci/joci
71
- - martn-nguyen/contrast_nli
72
- - pietrolesci/robust_nli
73
- - pietrolesci/robust_nli_is_sd
74
- - pietrolesci/robust_nli_li_ts
75
- - pietrolesci/gen_debiased_nli
76
- - pietrolesci/add_one_rte
77
- - metaeval/imppres
78
- - pietrolesci/glue_diagnostics
79
- - hlgd
80
- - PolyAI/banking77
81
- - paws
82
- - quora
83
- - medical_questions_pairs
84
- - conll2003
85
- - nlpaueb/finer-139
86
- - Anthropic/hh-rlhf
87
- - Anthropic/model-written-evals
88
- - truthful_qa
89
- - nightingal3/fig-qa
90
- - tasksource/bigbench
91
- - blimp
92
- - cos_e
93
- - cosmos_qa
94
- - dream
95
- - openbookqa
96
- - qasc
97
- - quartz
98
- - quail
99
- - head_qa
100
- - sciq
101
- - social_i_qa
102
- - wiki_hop
103
- - wiqa
104
- - piqa
105
- - hellaswag
106
- - pkavumba/balanced-copa
107
- - 12ml/e-CARE
108
- - art
109
- - tasksource/mmlu
110
- - winogrande
111
- - codah
112
- - ai2_arc
113
- - definite_pronoun_resolution
114
- - swag
115
- - math_qa
116
- - metaeval/utilitarianism
117
- - mteb/amazon_counterfactual
118
- - SetFit/insincere-questions
119
- - SetFit/toxic_conversations
120
- - turingbench/TuringBench
121
- - trec
122
- - tals/vitaminc
123
- - hope_edi
124
- - strombergnlp/rumoureval_2019
125
- - ethos
126
- - tweet_eval
127
- - discovery
128
- - pragmeval
129
- - silicone
130
- - lex_glue
131
- - papluca/language-identification
132
- - imdb
133
- - rotten_tomatoes
134
- - ag_news
135
- - yelp_review_full
136
- - financial_phrasebank
137
- - poem_sentiment
138
- - dbpedia_14
139
- - amazon_polarity
140
- - app_reviews
141
- - hate_speech18
142
- - sms_spam
143
- - humicroedit
144
- - snips_built_in_intents
145
- - banking77
146
- - hate_speech_offensive
147
- - yahoo_answers_topics
148
- - pacovaldez/stackoverflow-questions
149
- - zapsdcn/hyperpartisan_news
150
- - zapsdcn/sciie
151
- - zapsdcn/citation_intent
152
- - go_emotions
153
- - allenai/scicite
154
- - liar
155
- - relbert/lexical_relation_classification
156
- - metaeval/linguisticprobing
157
- - tasksource/crowdflower
158
- - metaeval/ethics
159
- - emo
160
- - google_wellformed_query
161
- - tweets_hate_speech_detection
162
- - has_part
163
- - wnut_17
164
- - ncbi_disease
165
- - acronym_identification
166
- - jnlpba
167
- - species_800
168
- - SpeedOfMagic/ontonotes_english
169
- - blog_authorship_corpus
170
- - launch/open_question_type
171
- - health_fact
172
- - commonsense_qa
173
- - mc_taco
174
- - ade_corpus_v2
175
- - prajjwal1/discosense
176
- - circa
177
- - PiC/phrase_similarity
178
- - copenlu/scientific-exaggeration-detection
179
- - quarel
180
- - mwong/fever-evidence-related
181
- - numer_sense
182
- - dynabench/dynasent
183
- - raquiba/Sarcasm_News_Headline
184
- - sem_eval_2010_task_8
185
- - demo-org/auditor_review
186
- - medmcqa
187
- - aqua_rat
188
- - RuyuanWan/Dynasent_Disagreement
189
- - RuyuanWan/Politeness_Disagreement
190
- - RuyuanWan/SBIC_Disagreement
191
- - RuyuanWan/SChem_Disagreement
192
- - RuyuanWan/Dilemmas_Disagreement
193
- - lucasmccabe/logiqa
194
- - wiki_qa
195
- - metaeval/cycic_classification
196
- - metaeval/cycic_multiplechoice
197
- - metaeval/sts-companion
198
- - metaeval/commonsense_qa_2.0
199
- - metaeval/lingnli
200
- - metaeval/monotonicity-entailment
201
- - metaeval/arct
202
- - metaeval/scinli
203
- - metaeval/naturallogic
204
- - onestop_qa
205
- - demelin/moral_stories
206
- - corypaik/prost
207
- - aps/dynahate
208
- - metaeval/syntactic-augmentation-nli
209
- - metaeval/autotnli
210
- - lasha-nlp/CONDAQA
211
- - openai/webgpt_comparisons
212
- - Dahoas/synthetic-instruct-gptj-pairwise
213
- - metaeval/scruples
214
- - metaeval/wouldyourather
215
- - sileod/attempto-nli
216
- - metaeval/defeasible-nli
217
- - metaeval/help-nli
218
- - metaeval/nli-veridicality-transitivity
219
- - metaeval/natural-language-satisfiability
220
- - metaeval/lonli
221
- - tasksource/dadc-limit-nli
222
- - ColumbiaNLP/FLUTE
223
- - metaeval/strategy-qa
224
- - openai/summarize_from_feedback
225
- - tasksource/folio
226
- - metaeval/tomi-nli
227
- - metaeval/avicenna
228
- - stanfordnlp/SHP
229
- - GBaker/MedQA-USMLE-4-options-hf
230
- - GBaker/MedQA-USMLE-4-options
231
- - sileod/wikimedqa
232
- - declare-lab/cicero
233
- - amydeng2000/CREAK
234
- - metaeval/mutual
235
- - inverse-scaling/NeQA
236
- - inverse-scaling/quote-repetition
237
- - inverse-scaling/redefine-math
238
- - tasksource/puzzte
239
- - metaeval/implicatures
240
- - race
241
- - metaeval/spartqa-yn
242
- - metaeval/spartqa-mchoice
243
- - metaeval/temporal-nli
244
- - metaeval/ScienceQA_text_only
245
- - AndyChiang/cloth
246
- - metaeval/logiqa-2.0-nli
247
- - tasksource/oasst1_dense_flat
248
- - metaeval/boolq-natural-perturbations
249
- - metaeval/path-naturalness-prediction
250
- - riddle_sense
251
- - Jiangjie/ekar_english
252
- - metaeval/implicit-hate-stg1
253
- - metaeval/chaos-mnli-ambiguity
254
- - IlyaGusev/headline_cause
255
- - metaeval/race-c
256
- - metaeval/equate
257
- - metaeval/ambient
258
- - AndyChiang/dgen
259
- - metaeval/clcd-english
260
- - civil_comments
261
- - metaeval/acceptability-prediction
262
- - maximedb/twentyquestions
263
- - metaeval/counterfactually-augmented-snli
264
- - tasksource/I2D2
265
- - sileod/mindgames
266
- - metaeval/counterfactually-augmented-imdb
267
- - metaeval/cnli
268
- - metaeval/reclor
269
- - tasksource/oasst1_pairwise_rlhf_reward
270
- - tasksource/zero-shot-label-nli
271
- - webis/args_me
272
- - webis/Touche23-ValueEval
273
- - tasksource/starcon
274
- - tasksource/ruletaker
275
- - lighteval/lsat_qa
276
- - tasksource/ConTRoL-nli
277
- - tasksource/tracie
278
- - tasksource/sherliic
279
- - tasksource/sen-making
280
- metrics:
281
- - accuracy
282
- library_name: transformers
283
- pipeline_tag: zero-shot-classification
284
- ---
285
-
286
  # Model Card for DeBERTa-v3-base-tasksource-nli
287
 
288
  This is [DeBERTa-v3-base](https://hf.co/microsoft/deberta-v3-base) fine-tuned with multi-task learning on 600 tasks.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  # Model Card for DeBERTa-v3-base-tasksource-nli
2
 
3
  This is [DeBERTa-v3-base](https://hf.co/microsoft/deberta-v3-base) fine-tuned with multi-task learning on 600 tasks.