albertvillanova HF staff commited on
Commit
0e05c2a
1 Parent(s): b7f730b

Add nli data files

Browse files
README.md CHANGED
@@ -150,13 +150,13 @@ dataset_info:
150
  '2': contradiction
151
  splits:
152
  - name: train
153
- num_bytes: 5719930
154
  num_examples: 24998
155
  - name: validation
156
- num_bytes: 673276
157
  num_examples: 3000
158
- download_size: 1257374
159
- dataset_size: 6393206
160
  - config_name: re
161
  features:
162
  - name: guid
@@ -313,6 +313,12 @@ dataset_info:
313
  download_size: 5012303
314
  dataset_size: 12148765
315
  configs:
 
 
 
 
 
 
316
  - config_name: sts
317
  data_files:
318
  - split: train
 
150
  '2': contradiction
151
  splits:
152
  - name: train
153
+ num_bytes: 5719882
154
  num_examples: 24998
155
  - name: validation
156
+ num_bytes: 673260
157
  num_examples: 3000
158
+ download_size: 2056116
159
+ dataset_size: 6393142
160
  - config_name: re
161
  features:
162
  - name: guid
 
313
  download_size: 5012303
314
  dataset_size: 12148765
315
  configs:
316
+ - config_name: nli
317
+ data_files:
318
+ - split: train
319
+ path: nli/train-*
320
+ - split: validation
321
+ path: nli/validation-*
322
  - config_name: sts
323
  data_files:
324
  - split: train
dataset_infos.json CHANGED
@@ -138,40 +138,31 @@
138
  "features": {
139
  "guid": {
140
  "dtype": "string",
141
- "id": null,
142
  "_type": "Value"
143
  },
144
  "source": {
145
  "dtype": "string",
146
- "id": null,
147
  "_type": "Value"
148
  },
149
  "premise": {
150
  "dtype": "string",
151
- "id": null,
152
  "_type": "Value"
153
  },
154
  "hypothesis": {
155
  "dtype": "string",
156
- "id": null,
157
  "_type": "Value"
158
  },
159
  "label": {
160
- "num_classes": 3,
161
  "names": [
162
  "entailment",
163
  "neutral",
164
  "contradiction"
165
  ],
166
- "names_file": null,
167
- "id": null,
168
  "_type": "ClassLabel"
169
  }
170
  },
171
- "post_processed": null,
172
- "supervised_keys": null,
173
- "task_templates": null,
174
  "builder_name": "klue",
 
175
  "config_name": "nli",
176
  "version": {
177
  "version_str": "1.0.0",
@@ -183,27 +174,20 @@
183
  "splits": {
184
  "train": {
185
  "name": "train",
186
- "num_bytes": 5719930,
187
  "num_examples": 24998,
188
- "dataset_name": "klue"
189
  },
190
  "validation": {
191
  "name": "validation",
192
- "num_bytes": 673276,
193
  "num_examples": 3000,
194
- "dataset_name": "klue"
195
- }
196
- },
197
- "download_checksums": {
198
- "http://klue-benchmark.com.s3.amazonaws.com/app/Competitions/000068/data/klue-nli-v1.tar.gz": {
199
- "num_bytes": 1257374,
200
- "checksum": "388be2033ef712072201903795a35b4f86826ee3ed3b62dc0c98e1721baa8850"
201
  }
202
  },
203
- "download_size": 1257374,
204
- "post_processing_size": null,
205
- "dataset_size": 6393206,
206
- "size_in_bytes": 7650580
207
  },
208
  "ner": {
209
  "description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n",
 
138
  "features": {
139
  "guid": {
140
  "dtype": "string",
 
141
  "_type": "Value"
142
  },
143
  "source": {
144
  "dtype": "string",
 
145
  "_type": "Value"
146
  },
147
  "premise": {
148
  "dtype": "string",
 
149
  "_type": "Value"
150
  },
151
  "hypothesis": {
152
  "dtype": "string",
 
153
  "_type": "Value"
154
  },
155
  "label": {
 
156
  "names": [
157
  "entailment",
158
  "neutral",
159
  "contradiction"
160
  ],
 
 
161
  "_type": "ClassLabel"
162
  }
163
  },
 
 
 
164
  "builder_name": "klue",
165
+ "dataset_name": "klue",
166
  "config_name": "nli",
167
  "version": {
168
  "version_str": "1.0.0",
 
174
  "splits": {
175
  "train": {
176
  "name": "train",
177
+ "num_bytes": 5719882,
178
  "num_examples": 24998,
179
+ "dataset_name": null
180
  },
181
  "validation": {
182
  "name": "validation",
183
+ "num_bytes": 673260,
184
  "num_examples": 3000,
185
+ "dataset_name": null
 
 
 
 
 
 
186
  }
187
  },
188
+ "download_size": 2056116,
189
+ "dataset_size": 6393142,
190
+ "size_in_bytes": 8449258
 
191
  },
192
  "ner": {
193
  "description": "KLUE (Korean Language Understanding Evaluation)\nKorean Language Understanding Evaluation (KLUE) benchmark is a series of datasets to evaluate natural language\nunderstanding capability of Korean language models. KLUE consists of 8 diverse and representative tasks, which are accessible\nto anyone without any restrictions. With ethical considerations in mind, we deliberately design annotation guidelines to obtain\nunambiguous annotations for all datasets. Futhermore, we build an evaluation system and carefully choose evaluations metrics\nfor every task, thus establishing fair comparison across Korean language models.\n",
nli/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35a8eb31818b94d9cfcb3477914360ce038b46b3bce4075eaef92309caa2c07c
3
+ size 1832150
nli/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b809f8f0aac2a2f41cd095bf15d8dbca9c931d7539390b6ef6241ca976195423
3
+ size 223966