Datasets:

Multilinguality:
multilingual
Size Categories:
1K<n<10K
Annotations Creators:
crowdsourced
ArXiv:
Tags:
License:
albertvillanova HF staff commited on
Commit
f9d9a27
1 Parent(s): 643641c

Add X-CSQA-hi data files

Browse files
README.md CHANGED
@@ -655,13 +655,13 @@ dataset_info:
655
  dtype: string
656
  splits:
657
  - name: test
658
- num_bytes: 415313
659
  num_examples: 1074
660
  - name: validation
661
- num_bytes: 396600
662
  num_examples: 1000
663
- download_size: 7519903
664
- dataset_size: 811913
665
  - config_name: X-CSQA-it
666
  features:
667
  - name: id
@@ -963,6 +963,12 @@ configs:
963
  path: X-CSQA-fr/test-*
964
  - split: validation
965
  path: X-CSQA-fr/validation-*
 
 
 
 
 
 
966
  - config_name: X-CSQA-it
967
  data_files:
968
  - split: test
 
655
  dtype: string
656
  splits:
657
  - name: test
658
+ num_bytes: 415011
659
  num_examples: 1074
660
  - name: validation
661
+ num_bytes: 396318
662
  num_examples: 1000
663
+ download_size: 304090
664
+ dataset_size: 811329
665
  - config_name: X-CSQA-it
666
  features:
667
  - name: id
 
963
  path: X-CSQA-fr/test-*
964
  - split: validation
965
  path: X-CSQA-fr/validation-*
966
+ - config_name: X-CSQA-hi
967
+ data_files:
968
+ - split: test
969
+ path: X-CSQA-hi/test-*
970
+ - split: validation
971
+ path: X-CSQA-hi/validation-*
972
  - config_name: X-CSQA-it
973
  data_files:
974
  - split: test
X-CSQA-hi/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3642eef63f363ec2c25ef48e4d20a508f9002303f17852516b4b93f67717b60d
3
+ size 157568
X-CSQA-hi/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd4cccd74e03bb58e38b6a138832e0c29785bac3781592047eb5563391fd066c
3
+ size 146522
dataset_infos.json CHANGED
@@ -865,48 +865,38 @@
865
  "features": {
866
  "id": {
867
  "dtype": "string",
868
- "id": null,
869
  "_type": "Value"
870
  },
871
  "lang": {
872
  "dtype": "string",
873
- "id": null,
874
  "_type": "Value"
875
  },
876
  "question": {
877
  "stem": {
878
  "dtype": "string",
879
- "id": null,
880
  "_type": "Value"
881
  },
882
  "choices": {
883
  "feature": {
884
  "label": {
885
  "dtype": "string",
886
- "id": null,
887
  "_type": "Value"
888
  },
889
  "text": {
890
  "dtype": "string",
891
- "id": null,
892
  "_type": "Value"
893
  }
894
  },
895
- "length": -1,
896
- "id": null,
897
  "_type": "Sequence"
898
  }
899
  },
900
  "answerKey": {
901
  "dtype": "string",
902
- "id": null,
903
  "_type": "Value"
904
  }
905
  },
906
- "post_processed": null,
907
- "supervised_keys": null,
908
- "task_templates": null,
909
  "builder_name": "xcsr",
 
910
  "config_name": "X-CSQA-hi",
911
  "version": {
912
  "version_str": "1.1.0",
@@ -918,27 +908,20 @@
918
  "splits": {
919
  "test": {
920
  "name": "test",
921
- "num_bytes": 415313,
922
  "num_examples": 1074,
923
- "dataset_name": "xcsr"
924
  },
925
  "validation": {
926
  "name": "validation",
927
- "num_bytes": 396600,
928
  "num_examples": 1000,
929
- "dataset_name": "xcsr"
930
- }
931
- },
932
- "download_checksums": {
933
- "https://inklab.usc.edu/XCSR/xcsr_datasets.zip": {
934
- "num_bytes": 7519903,
935
- "checksum": "c45b29ece740643252d5402e76be1e33f96f9d6910053f79e80d39887f10c85e"
936
  }
937
  },
938
- "download_size": 7519903,
939
- "post_processing_size": null,
940
- "dataset_size": 811913,
941
- "size_in_bytes": 8331816
942
  },
943
  "X-CSQA-sw": {
944
  "description": "To evaluate multi-lingual language models (ML-LMs) for commonsense reasoning in a cross-lingual zero-shot transfer setting (X-CSR), i.e., training in English and test in other languages, we create two benchmark datasets, namely X-CSQA and X-CODAH. Specifically, we automatically translate the original CSQA and CODAH datasets, which only have English versions, to 15 other languages, forming development and test sets for studying X-CSR. As our goal is to evaluate different ML-LMs in a unified evaluation protocol for X-CSR, we argue that such translated examples, although might contain noise, can serve as a starting benchmark for us to obtain meaningful analysis, before more human-translated datasets will be available in the future.\n",
 
865
  "features": {
866
  "id": {
867
  "dtype": "string",
 
868
  "_type": "Value"
869
  },
870
  "lang": {
871
  "dtype": "string",
 
872
  "_type": "Value"
873
  },
874
  "question": {
875
  "stem": {
876
  "dtype": "string",
 
877
  "_type": "Value"
878
  },
879
  "choices": {
880
  "feature": {
881
  "label": {
882
  "dtype": "string",
 
883
  "_type": "Value"
884
  },
885
  "text": {
886
  "dtype": "string",
 
887
  "_type": "Value"
888
  }
889
  },
 
 
890
  "_type": "Sequence"
891
  }
892
  },
893
  "answerKey": {
894
  "dtype": "string",
 
895
  "_type": "Value"
896
  }
897
  },
 
 
 
898
  "builder_name": "xcsr",
899
+ "dataset_name": "xcsr",
900
  "config_name": "X-CSQA-hi",
901
  "version": {
902
  "version_str": "1.1.0",
 
908
  "splits": {
909
  "test": {
910
  "name": "test",
911
+ "num_bytes": 415011,
912
  "num_examples": 1074,
913
+ "dataset_name": null
914
  },
915
  "validation": {
916
  "name": "validation",
917
+ "num_bytes": 396318,
918
  "num_examples": 1000,
919
+ "dataset_name": null
 
 
 
 
 
 
920
  }
921
  },
922
+ "download_size": 304090,
923
+ "dataset_size": 811329,
924
+ "size_in_bytes": 1115419
 
925
  },
926
  "X-CSQA-sw": {
927
  "description": "To evaluate multi-lingual language models (ML-LMs) for commonsense reasoning in a cross-lingual zero-shot transfer setting (X-CSR), i.e., training in English and test in other languages, we create two benchmark datasets, namely X-CSQA and X-CODAH. Specifically, we automatically translate the original CSQA and CODAH datasets, which only have English versions, to 15 other languages, forming development and test sets for studying X-CSR. As our goal is to evaluate different ML-LMs in a unified evaluation protocol for X-CSR, we argue that such translated examples, although might contain noise, can serve as a starting benchmark for us to obtain meaningful analysis, before more human-translated datasets will be available in the future.\n",