Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:
albertvillanova HF staff commited on
Commit
ffd96bd
1 Parent(s): 210a4d2

Convert dataset to Parquet

Browse files

Convert dataset to Parquet.

README.md CHANGED
@@ -1,5 +1,4 @@
1
  ---
2
- pretty_name: DROP
3
  annotations_creators:
4
  - crowdsourced
5
  language_creators:
@@ -21,6 +20,7 @@ task_ids:
21
  - extractive-qa
22
  - abstractive-qa
23
  paperswithcode_id: drop
 
24
  dataset_info:
25
  features:
26
  - name: section_id
@@ -39,13 +39,20 @@ dataset_info:
39
  dtype: string
40
  splits:
41
  - name: train
42
- num_bytes: 105572762
43
  num_examples: 77400
44
  - name: validation
45
- num_bytes: 11737787
46
  num_examples: 9535
47
- download_size: 8308692
48
- dataset_size: 117310549
 
 
 
 
 
 
 
49
  ---
50
 
51
  # Dataset Card for "drop"
 
1
  ---
 
2
  annotations_creators:
3
  - crowdsourced
4
  language_creators:
 
20
  - extractive-qa
21
  - abstractive-qa
22
  paperswithcode_id: drop
23
+ pretty_name: DROP
24
  dataset_info:
25
  features:
26
  - name: section_id
 
39
  dtype: string
40
  splits:
41
  - name: train
42
+ num_bytes: 105572506
43
  num_examples: 77400
44
  - name: validation
45
+ num_bytes: 11737755
46
  num_examples: 9535
47
+ download_size: 11538387
48
+ dataset_size: 117310261
49
+ configs:
50
+ - config_name: default
51
+ data_files:
52
+ - split: train
53
+ path: data/train-*
54
+ - split: validation
55
+ path: data/validation-*
56
  ---
57
 
58
  # Dataset Card for "drop"
data/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abb5e578c2156a61f83b56c066a922ec1c7c5140638a3f0f2a7c348fafe1cb35
3
+ size 10333127
data/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f9a3bdbb1b5909abfa25cbab693f89f47568c98e6e03473500d604f044c8f68
3
+ size 1205260
dataset_infos.json CHANGED
@@ -1 +1,65 @@
1
- {"default": {"description": "DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs.\n. DROP is a crowdsourced, adversarially-created, 96k-question benchmark, in which a system must resolve references in a\nquestion, perhaps to multiple input positions, and perform discrete operations over them (such as addition, counting, or\n sorting). These operations require a much more comprehensive understanding of the content of paragraphs than what was\n necessary for prior datasets.\n", "citation": "@inproceedings{Dua2019DROP,\n author={Dheeru Dua and Yizhong Wang and Pradeep Dasigi and Gabriel Stanovsky and Sameer Singh and Matt Gardner},\n title={DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs},\n booktitle={Proc. of NAACL},\n year={2019}\n}\n", "homepage": "https://allennlp.org/drop", "license": "", "features": {"section_id": {"dtype": "string", "id": null, "_type": "Value"}, "query_id": {"dtype": "string", "id": null, "_type": "Value"}, "passage": {"dtype": "string", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "answers_spans": {"feature": {"spans": {"dtype": "string", "id": null, "_type": "Value"}, "types": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "builder_name": "drop", "config_name": "default", "version": {"version_str": "0.1.0", "description": null, "major": 0, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 105572762, "num_examples": 77400, "dataset_name": "drop"}, "validation": {"name": "validation", "num_bytes": 11737787, "num_examples": 9535, "dataset_name": "drop"}}, "download_checksums": {"https://s3-us-west-2.amazonaws.com/allennlp/datasets/drop/drop_dataset.zip": {"num_bytes": 8308692, "checksum": "39d2278a29fd729de301b111a45f434c24834f40df8f4ff116d864589e3249d6"}}, "download_size": 8308692, "post_processing_size": null, "dataset_size": 117310549, "size_in_bytes": 125619241}}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "default": {
3
+ "description": "DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs.\n. DROP is a crowdsourced, adversarially-created, 96k-question benchmark, in which a system must resolve references in a\nquestion, perhaps to multiple input positions, and perform discrete operations over them (such as addition, counting, or\n sorting). These operations require a much more comprehensive understanding of the content of paragraphs than what was\n necessary for prior datasets.\n",
4
+ "citation": "@inproceedings{Dua2019DROP,\n author={Dheeru Dua and Yizhong Wang and Pradeep Dasigi and Gabriel Stanovsky and Sameer Singh and Matt Gardner},\n title={DROP: A Reading Comprehension Benchmark Requiring Discrete Reasoning Over Paragraphs},\n booktitle={Proc. of NAACL},\n year={2019}\n}\n",
5
+ "homepage": "https://allennlp.org/drop",
6
+ "license": "",
7
+ "features": {
8
+ "section_id": {
9
+ "dtype": "string",
10
+ "_type": "Value"
11
+ },
12
+ "query_id": {
13
+ "dtype": "string",
14
+ "_type": "Value"
15
+ },
16
+ "passage": {
17
+ "dtype": "string",
18
+ "_type": "Value"
19
+ },
20
+ "question": {
21
+ "dtype": "string",
22
+ "_type": "Value"
23
+ },
24
+ "answers_spans": {
25
+ "feature": {
26
+ "spans": {
27
+ "dtype": "string",
28
+ "_type": "Value"
29
+ },
30
+ "types": {
31
+ "dtype": "string",
32
+ "_type": "Value"
33
+ }
34
+ },
35
+ "_type": "Sequence"
36
+ }
37
+ },
38
+ "builder_name": "drop",
39
+ "dataset_name": "drop",
40
+ "config_name": "default",
41
+ "version": {
42
+ "version_str": "0.1.0",
43
+ "major": 0,
44
+ "minor": 1,
45
+ "patch": 0
46
+ },
47
+ "splits": {
48
+ "train": {
49
+ "name": "train",
50
+ "num_bytes": 105572506,
51
+ "num_examples": 77400,
52
+ "dataset_name": null
53
+ },
54
+ "validation": {
55
+ "name": "validation",
56
+ "num_bytes": 11737755,
57
+ "num_examples": 9535,
58
+ "dataset_name": null
59
+ }
60
+ },
61
+ "download_size": 11538387,
62
+ "dataset_size": 117310261,
63
+ "size_in_bytes": 128848648
64
+ }
65
+ }