albertvillanova HF staff commited on
Commit
554ee79
1 Parent(s): 507a95b

Add adv_mnli_mismatched data files

Browse files
README.md CHANGED
@@ -68,10 +68,10 @@ dataset_info:
68
  dtype: int32
69
  splits:
70
  - name: validation
71
- num_bytes: 40982
72
  num_examples: 162
73
- download_size: 40662
74
- dataset_size: 40982
75
  - config_name: adv_qnli
76
  features:
77
  - name: question
@@ -155,6 +155,10 @@ configs:
155
  data_files:
156
  - split: validation
157
  path: adv_mnli/validation-*
 
 
 
 
158
  ---
159
 
160
  # Dataset Card for Adversarial GLUE
 
68
  dtype: int32
69
  splits:
70
  - name: validation
71
+ num_bytes: 40953
72
  num_examples: 162
73
+ download_size: 25166
74
+ dataset_size: 40953
75
  - config_name: adv_qnli
76
  features:
77
  - name: question
 
155
  data_files:
156
  - split: validation
157
  path: adv_mnli/validation-*
158
+ - config_name: adv_mnli_mismatched
159
+ data_files:
160
+ - split: validation
161
+ path: adv_mnli_mismatched/validation-*
162
  ---
163
 
164
  # Dataset Card for Adversarial GLUE
adv_mnli_mismatched/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c65be9dc6735d5c1a9c4ec9b3fe6e822839d28d6da8707c2a1b2ad217b8bcf12
3
+ size 25166
dataset_infos.json CHANGED
@@ -174,38 +174,30 @@
174
  "features": {
175
  "premise": {
176
  "dtype": "string",
177
- "id": null,
178
  "_type": "Value"
179
  },
180
  "hypothesis": {
181
  "dtype": "string",
182
- "id": null,
183
  "_type": "Value"
184
  },
185
  "label": {
186
- "num_classes": 3,
187
  "names": [
188
  "entailment",
189
  "neutral",
190
  "contradiction"
191
  ],
192
- "id": null,
193
  "_type": "ClassLabel"
194
  },
195
  "idx": {
196
  "dtype": "int32",
197
- "id": null,
198
  "_type": "Value"
199
  }
200
  },
201
- "post_processed": null,
202
- "supervised_keys": null,
203
- "task_templates": null,
204
- "builder_name": "adv_glue",
205
  "config_name": "adv_mnli_mismatched",
206
  "version": {
207
  "version_str": "1.0.0",
208
- "description": "",
209
  "major": 1,
210
  "minor": 0,
211
  "patch": 0
@@ -213,21 +205,14 @@
213
  "splits": {
214
  "validation": {
215
  "name": "validation",
216
- "num_bytes": 40982,
217
  "num_examples": 162,
218
- "dataset_name": "adv_glue"
219
- }
220
- },
221
- "download_checksums": {
222
- "https://adversarialglue.github.io/dataset/dev.zip": {
223
- "num_bytes": 40662,
224
- "checksum": "efb4cbd3aa4a87bfaffc310ae951981cc0a36c6c71c6425dd74e5b55f2f325c9"
225
  }
226
  },
227
- "download_size": 40662,
228
- "post_processing_size": null,
229
- "dataset_size": 40982,
230
- "size_in_bytes": 81644
231
  },
232
  "adv_qnli": {
233
  "description": "Adversarial GLUE Benchmark (AdvGLUE) is a comprehensive robustness evaluation benchmark\nthat focuses on the adversarial robustness evaluation of language models. It covers five\nnatural language understanding tasks from the famous GLUE tasks and is an adversarial\nversion of GLUE benchmark.\n",
 
174
  "features": {
175
  "premise": {
176
  "dtype": "string",
 
177
  "_type": "Value"
178
  },
179
  "hypothesis": {
180
  "dtype": "string",
 
181
  "_type": "Value"
182
  },
183
  "label": {
 
184
  "names": [
185
  "entailment",
186
  "neutral",
187
  "contradiction"
188
  ],
 
189
  "_type": "ClassLabel"
190
  },
191
  "idx": {
192
  "dtype": "int32",
 
193
  "_type": "Value"
194
  }
195
  },
196
+ "builder_name": "parquet",
197
+ "dataset_name": "adv_glue",
 
 
198
  "config_name": "adv_mnli_mismatched",
199
  "version": {
200
  "version_str": "1.0.0",
 
201
  "major": 1,
202
  "minor": 0,
203
  "patch": 0
 
205
  "splits": {
206
  "validation": {
207
  "name": "validation",
208
+ "num_bytes": 40953,
209
  "num_examples": 162,
210
+ "dataset_name": null
 
 
 
 
 
 
211
  }
212
  },
213
+ "download_size": 25166,
214
+ "dataset_size": 40953,
215
+ "size_in_bytes": 66119
 
216
  },
217
  "adv_qnli": {
218
  "description": "Adversarial GLUE Benchmark (AdvGLUE) is a comprehensive robustness evaluation benchmark\nthat focuses on the adversarial robustness evaluation of language models. It covers five\nnatural language understanding tasks from the famous GLUE tasks and is an adversarial\nversion of GLUE benchmark.\n",