Datasets:

Multilinguality:
multilingual
Size Categories:
1K<n<10K
Annotations Creators:
crowdsourced
ArXiv:
Tags:
License:
albertvillanova HF staff commited on
Commit
e9d7276
1 Parent(s): 63a5ce5

Add X-CSQA-nl data files

Browse files
README.md CHANGED
@@ -736,13 +736,13 @@ dataset_info:
736
  dtype: string
737
  splits:
738
  - name: test
739
- num_bytes: 227251
740
  num_examples: 1074
741
  - name: validation
742
- num_bytes: 216476
743
  num_examples: 1000
744
- download_size: 7519903
745
- dataset_size: 443727
746
  - config_name: X-CSQA-pl
747
  features:
748
  - name: id
@@ -969,6 +969,12 @@ configs:
969
  path: X-CSQA-jap/test-*
970
  - split: validation
971
  path: X-CSQA-jap/validation-*
 
 
 
 
 
 
972
  - config_name: X-CSQA-zh
973
  data_files:
974
  - split: test
 
736
  dtype: string
737
  splits:
738
  - name: test
739
+ num_bytes: 226949
740
  num_examples: 1074
741
  - name: validation
742
+ num_bytes: 216194
743
  num_examples: 1000
744
+ download_size: 231078
745
+ dataset_size: 443143
746
  - config_name: X-CSQA-pl
747
  features:
748
  - name: id
 
969
  path: X-CSQA-jap/test-*
970
  - split: validation
971
  path: X-CSQA-jap/validation-*
972
+ - config_name: X-CSQA-nl
973
+ data_files:
974
+ - split: test
975
+ path: X-CSQA-nl/test-*
976
+ - split: validation
977
+ path: X-CSQA-nl/validation-*
978
  - config_name: X-CSQA-zh
979
  data_files:
980
  - split: test
X-CSQA-nl/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8aa3569286952a0114fa7ec9f7f27625dd9005ea7c96ffc1bd9f9c4ca96c2303
3
+ size 119325
X-CSQA-nl/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91865e7fc15d5b081a08f3f9ef0b42284e3f65600da81738649db978af013f1b
3
+ size 111753
dataset_infos.json CHANGED
@@ -469,48 +469,38 @@
469
  "features": {
470
  "id": {
471
  "dtype": "string",
472
- "id": null,
473
  "_type": "Value"
474
  },
475
  "lang": {
476
  "dtype": "string",
477
- "id": null,
478
  "_type": "Value"
479
  },
480
  "question": {
481
  "stem": {
482
  "dtype": "string",
483
- "id": null,
484
  "_type": "Value"
485
  },
486
  "choices": {
487
  "feature": {
488
  "label": {
489
  "dtype": "string",
490
- "id": null,
491
  "_type": "Value"
492
  },
493
  "text": {
494
  "dtype": "string",
495
- "id": null,
496
  "_type": "Value"
497
  }
498
  },
499
- "length": -1,
500
- "id": null,
501
  "_type": "Sequence"
502
  }
503
  },
504
  "answerKey": {
505
  "dtype": "string",
506
- "id": null,
507
  "_type": "Value"
508
  }
509
  },
510
- "post_processed": null,
511
- "supervised_keys": null,
512
- "task_templates": null,
513
  "builder_name": "xcsr",
 
514
  "config_name": "X-CSQA-nl",
515
  "version": {
516
  "version_str": "1.1.0",
@@ -522,27 +512,20 @@
522
  "splits": {
523
  "test": {
524
  "name": "test",
525
- "num_bytes": 227251,
526
  "num_examples": 1074,
527
- "dataset_name": "xcsr"
528
  },
529
  "validation": {
530
  "name": "validation",
531
- "num_bytes": 216476,
532
  "num_examples": 1000,
533
- "dataset_name": "xcsr"
534
- }
535
- },
536
- "download_checksums": {
537
- "https://inklab.usc.edu/XCSR/xcsr_datasets.zip": {
538
- "num_bytes": 7519903,
539
- "checksum": "c45b29ece740643252d5402e76be1e33f96f9d6910053f79e80d39887f10c85e"
540
  }
541
  },
542
- "download_size": 7519903,
543
- "post_processing_size": null,
544
- "dataset_size": 443727,
545
- "size_in_bytes": 7963630
546
  },
547
  "X-CSQA-pl": {
548
  "description": "To evaluate multi-lingual language models (ML-LMs) for commonsense reasoning in a cross-lingual zero-shot transfer setting (X-CSR), i.e., training in English and test in other languages, we create two benchmark datasets, namely X-CSQA and X-CODAH. Specifically, we automatically translate the original CSQA and CODAH datasets, which only have English versions, to 15 other languages, forming development and test sets for studying X-CSR. As our goal is to evaluate different ML-LMs in a unified evaluation protocol for X-CSR, we argue that such translated examples, although might contain noise, can serve as a starting benchmark for us to obtain meaningful analysis, before more human-translated datasets will be available in the future.\n",
 
469
  "features": {
470
  "id": {
471
  "dtype": "string",
 
472
  "_type": "Value"
473
  },
474
  "lang": {
475
  "dtype": "string",
 
476
  "_type": "Value"
477
  },
478
  "question": {
479
  "stem": {
480
  "dtype": "string",
 
481
  "_type": "Value"
482
  },
483
  "choices": {
484
  "feature": {
485
  "label": {
486
  "dtype": "string",
 
487
  "_type": "Value"
488
  },
489
  "text": {
490
  "dtype": "string",
 
491
  "_type": "Value"
492
  }
493
  },
 
 
494
  "_type": "Sequence"
495
  }
496
  },
497
  "answerKey": {
498
  "dtype": "string",
 
499
  "_type": "Value"
500
  }
501
  },
 
 
 
502
  "builder_name": "xcsr",
503
+ "dataset_name": "xcsr",
504
  "config_name": "X-CSQA-nl",
505
  "version": {
506
  "version_str": "1.1.0",
 
512
  "splits": {
513
  "test": {
514
  "name": "test",
515
+ "num_bytes": 226949,
516
  "num_examples": 1074,
517
+ "dataset_name": null
518
  },
519
  "validation": {
520
  "name": "validation",
521
+ "num_bytes": 216194,
522
  "num_examples": 1000,
523
+ "dataset_name": null
 
 
 
 
 
 
524
  }
525
  },
526
+ "download_size": 231078,
527
+ "dataset_size": 443143,
528
+ "size_in_bytes": 674221
 
529
  },
530
  "X-CSQA-pl": {
531
  "description": "To evaluate multi-lingual language models (ML-LMs) for commonsense reasoning in a cross-lingual zero-shot transfer setting (X-CSR), i.e., training in English and test in other languages, we create two benchmark datasets, namely X-CSQA and X-CODAH. Specifically, we automatically translate the original CSQA and CODAH datasets, which only have English versions, to 15 other languages, forming development and test sets for studying X-CSR. As our goal is to evaluate different ML-LMs in a unified evaluation protocol for X-CSR, we argue that such translated examples, although might contain noise, can serve as a starting benchmark for us to obtain meaningful analysis, before more human-translated datasets will be available in the future.\n",