Datasets:

Multilinguality:
multilingual
Size Categories:
1K<n<10K
Annotations Creators:
crowdsourced
ArXiv:
Tags:
License:
albertvillanova HF staff commited on
Commit
4595814
1 Parent(s): 8a0429e

Add X-CODAH-de data files

Browse files
README.md CHANGED
@@ -87,13 +87,13 @@ dataset_info:
87
  dtype: string
88
  splits:
89
  - name: test
90
- num_bytes: 476373
91
  num_examples: 1000
92
  - name: validation
93
- num_bytes: 138876
94
  num_examples: 300
95
- download_size: 7519903
96
- dataset_size: 615249
97
  - config_name: X-CODAH-en
98
  features:
99
  - name: id
@@ -933,6 +933,12 @@ dataset_info:
933
  download_size: 207379
934
  dataset_size: 385717
935
  configs:
 
 
 
 
 
 
936
  - config_name: X-CODAH-en
937
  data_files:
938
  - split: test
 
87
  dtype: string
88
  splits:
89
  - name: test
90
+ num_bytes: 476087
91
  num_examples: 1000
92
  - name: validation
93
+ num_bytes: 138764
94
  num_examples: 300
95
+ download_size: 259705
96
+ dataset_size: 614851
97
  - config_name: X-CODAH-en
98
  features:
99
  - name: id
 
933
  download_size: 207379
934
  dataset_size: 385717
935
  configs:
936
+ - config_name: X-CODAH-de
937
+ data_files:
938
+ - split: test
939
+ path: X-CODAH-de/test-*
940
+ - split: validation
941
+ path: X-CODAH-de/validation-*
942
  - config_name: X-CODAH-en
943
  data_files:
944
  - split: test
X-CODAH-de/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebd0c6fb01472a92bef9c3e0f7bef493df689489797aeebd0286d3c0c19c4bc4
3
+ size 198977
X-CODAH-de/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3440067dd19d8fc908af8a166df3f08661ba190a0a72d1eb282d614eafa71312
3
+ size 60728
dataset_infos.json CHANGED
@@ -1203,53 +1203,42 @@
1203
  "features": {
1204
  "id": {
1205
  "dtype": "string",
1206
- "id": null,
1207
  "_type": "Value"
1208
  },
1209
  "lang": {
1210
  "dtype": "string",
1211
- "id": null,
1212
  "_type": "Value"
1213
  },
1214
  "question_tag": {
1215
  "dtype": "string",
1216
- "id": null,
1217
  "_type": "Value"
1218
  },
1219
  "question": {
1220
  "stem": {
1221
  "dtype": "string",
1222
- "id": null,
1223
  "_type": "Value"
1224
  },
1225
  "choices": {
1226
  "feature": {
1227
  "label": {
1228
  "dtype": "string",
1229
- "id": null,
1230
  "_type": "Value"
1231
  },
1232
  "text": {
1233
  "dtype": "string",
1234
- "id": null,
1235
  "_type": "Value"
1236
  }
1237
  },
1238
- "length": -1,
1239
- "id": null,
1240
  "_type": "Sequence"
1241
  }
1242
  },
1243
  "answerKey": {
1244
  "dtype": "string",
1245
- "id": null,
1246
  "_type": "Value"
1247
  }
1248
  },
1249
- "post_processed": null,
1250
- "supervised_keys": null,
1251
- "task_templates": null,
1252
  "builder_name": "xcsr",
 
1253
  "config_name": "X-CODAH-de",
1254
  "version": {
1255
  "version_str": "1.1.0",
@@ -1261,27 +1250,20 @@
1261
  "splits": {
1262
  "test": {
1263
  "name": "test",
1264
- "num_bytes": 476373,
1265
  "num_examples": 1000,
1266
- "dataset_name": "xcsr"
1267
  },
1268
  "validation": {
1269
  "name": "validation",
1270
- "num_bytes": 138876,
1271
  "num_examples": 300,
1272
- "dataset_name": "xcsr"
1273
- }
1274
- },
1275
- "download_checksums": {
1276
- "https://inklab.usc.edu/XCSR/xcsr_datasets.zip": {
1277
- "num_bytes": 7519903,
1278
- "checksum": "c45b29ece740643252d5402e76be1e33f96f9d6910053f79e80d39887f10c85e"
1279
  }
1280
  },
1281
- "download_size": 7519903,
1282
- "post_processing_size": null,
1283
- "dataset_size": 615249,
1284
- "size_in_bytes": 8135152
1285
  },
1286
  "X-CODAH-es": {
1287
  "description": "To evaluate multi-lingual language models (ML-LMs) for commonsense reasoning in a cross-lingual zero-shot transfer setting (X-CSR), i.e., training in English and test in other languages, we create two benchmark datasets, namely X-CSQA and X-CODAH. Specifically, we automatically translate the original CSQA and CODAH datasets, which only have English versions, to 15 other languages, forming development and test sets for studying X-CSR. As our goal is to evaluate different ML-LMs in a unified evaluation protocol for X-CSR, we argue that such translated examples, although might contain noise, can serve as a starting benchmark for us to obtain meaningful analysis, before more human-translated datasets will be available in the future.\n",
 
1203
  "features": {
1204
  "id": {
1205
  "dtype": "string",
 
1206
  "_type": "Value"
1207
  },
1208
  "lang": {
1209
  "dtype": "string",
 
1210
  "_type": "Value"
1211
  },
1212
  "question_tag": {
1213
  "dtype": "string",
 
1214
  "_type": "Value"
1215
  },
1216
  "question": {
1217
  "stem": {
1218
  "dtype": "string",
 
1219
  "_type": "Value"
1220
  },
1221
  "choices": {
1222
  "feature": {
1223
  "label": {
1224
  "dtype": "string",
 
1225
  "_type": "Value"
1226
  },
1227
  "text": {
1228
  "dtype": "string",
 
1229
  "_type": "Value"
1230
  }
1231
  },
 
 
1232
  "_type": "Sequence"
1233
  }
1234
  },
1235
  "answerKey": {
1236
  "dtype": "string",
 
1237
  "_type": "Value"
1238
  }
1239
  },
 
 
 
1240
  "builder_name": "xcsr",
1241
+ "dataset_name": "xcsr",
1242
  "config_name": "X-CODAH-de",
1243
  "version": {
1244
  "version_str": "1.1.0",
 
1250
  "splits": {
1251
  "test": {
1252
  "name": "test",
1253
+ "num_bytes": 476087,
1254
  "num_examples": 1000,
1255
+ "dataset_name": null
1256
  },
1257
  "validation": {
1258
  "name": "validation",
1259
+ "num_bytes": 138764,
1260
  "num_examples": 300,
1261
+ "dataset_name": null
 
 
 
 
 
 
1262
  }
1263
  },
1264
+ "download_size": 259705,
1265
+ "dataset_size": 614851,
1266
+ "size_in_bytes": 874556
 
1267
  },
1268
  "X-CODAH-es": {
1269
  "description": "To evaluate multi-lingual language models (ML-LMs) for commonsense reasoning in a cross-lingual zero-shot transfer setting (X-CSR), i.e., training in English and test in other languages, we create two benchmark datasets, namely X-CSQA and X-CODAH. Specifically, we automatically translate the original CSQA and CODAH datasets, which only have English versions, to 15 other languages, forming development and test sets for studying X-CSR. As our goal is to evaluate different ML-LMs in a unified evaluation protocol for X-CSR, we argue that such translated examples, although might contain noise, can serve as a starting benchmark for us to obtain meaningful analysis, before more human-translated datasets will be available in the future.\n",