Datasets:

Modalities:
Text
Formats:
parquet
Languages:
English
ArXiv:
Tags:
License:
albertvillanova HF staff commited on
Commit
72112ee
1 Parent(s): 4b3c61e

Convert dataset to Parquet

Browse files

Convert dataset to Parquet.

README.md CHANGED
@@ -21,7 +21,7 @@ task_ids:
21
  paperswithcode_id: ambigqa
22
  pretty_name: 'AmbigQA: Answering Ambiguous Open-domain Questions'
23
  dataset_info:
24
- - config_name: light
25
  features:
26
  - name: id
27
  dtype: string
@@ -39,16 +39,32 @@ dataset_info:
39
  dtype: string
40
  - name: answer
41
  sequence: string
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
42
  splits:
43
  - name: train
44
- num_bytes: 2739732
45
  num_examples: 10036
46
  - name: validation
47
- num_bytes: 805808
48
  num_examples: 2002
49
- download_size: 19700900
50
- dataset_size: 3545540
51
- - config_name: full
52
  features:
53
  - name: id
54
  dtype: string
@@ -66,31 +82,23 @@ dataset_info:
66
  dtype: string
67
  - name: answer
68
  sequence: string
69
- - name: viewed_doc_titles
70
- sequence: string
71
- - name: used_queries
72
- sequence:
73
- - name: query
74
- dtype: string
75
- - name: results
76
- sequence:
77
- - name: title
78
- dtype: string
79
- - name: snippet
80
- dtype: string
81
- - name: nq_answer
82
- sequence: string
83
- - name: nq_doc_title
84
- dtype: string
85
  splits:
86
  - name: train
87
- num_bytes: 43538733
88
  num_examples: 10036
89
  - name: validation
90
- num_bytes: 15383368
91
  num_examples: 2002
92
  download_size: 19700900
93
- dataset_size: 58922101
 
 
 
 
 
 
 
 
94
  ---
95
 
96
  # Dataset Card for AmbigQA: Answering Ambiguous Open-domain Questions
 
21
  paperswithcode_id: ambigqa
22
  pretty_name: 'AmbigQA: Answering Ambiguous Open-domain Questions'
23
  dataset_info:
24
+ - config_name: full
25
  features:
26
  - name: id
27
  dtype: string
 
39
  dtype: string
40
  - name: answer
41
  sequence: string
42
+ - name: viewed_doc_titles
43
+ sequence: string
44
+ - name: used_queries
45
+ sequence:
46
+ - name: query
47
+ dtype: string
48
+ - name: results
49
+ sequence:
50
+ - name: title
51
+ dtype: string
52
+ - name: snippet
53
+ dtype: string
54
+ - name: nq_answer
55
+ sequence: string
56
+ - name: nq_doc_title
57
+ dtype: string
58
  splits:
59
  - name: train
60
+ num_bytes: 43538533
61
  num_examples: 10036
62
  - name: validation
63
+ num_bytes: 15383268
64
  num_examples: 2002
65
+ download_size: 30674462
66
+ dataset_size: 58921801
67
+ - config_name: light
68
  features:
69
  - name: id
70
  dtype: string
 
82
  dtype: string
83
  - name: answer
84
  sequence: string
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
85
  splits:
86
  - name: train
87
+ num_bytes: 2739732
88
  num_examples: 10036
89
  - name: validation
90
+ num_bytes: 805808
91
  num_examples: 2002
92
  download_size: 19700900
93
+ dataset_size: 3545540
94
+ configs:
95
+ - config_name: full
96
+ data_files:
97
+ - split: train
98
+ path: full/train-*
99
+ - split: validation
100
+ path: full/validation-*
101
+ default: true
102
  ---
103
 
104
  # Dataset Card for AmbigQA: Answering Ambiguous Open-domain Questions
dataset_infos.json CHANGED
@@ -1 +1,217 @@
1
- {"light": {"description": "AmbigNQ, a dataset covering 14,042 questions from NQ-open, an existing open-domain QA benchmark. We find that over half of the questions in NQ-open are ambiguous. The types of ambiguity are diverse and sometimes subtle, many of which are only apparent after examining evidence provided by a very large text corpus. AMBIGNQ, a dataset with\n14,042 annotations on NQ-OPEN questions containing diverse types of ambiguity.\nWe provide two distributions of our new dataset AmbigNQ: a full version with all annotation metadata and a light version with only inputs and outputs.\n", "citation": "@inproceedings{ min2020ambigqa,\n title={ {A}mbig{QA}: Answering Ambiguous Open-domain Questions },\n author={ Min, Sewon and Michael, Julian and Hajishirzi, Hannaneh and Zettlemoyer, Luke },\n booktitle={ EMNLP },\n year={2020}\n}\n", "homepage": "https://nlp.cs.washington.edu/ambigqa/", "license": "CC BY-SA 3.0", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "annotations": {"feature": {"type": {"dtype": "string", "id": null, "_type": "Value"}, "answer": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "qaPairs": {"feature": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "answer": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}}, "length": -1, "id": null, "_type": "Sequence"}}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "builder_name": "ambig_qa", "config_name": "light", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 2739732, "num_examples": 10036, "dataset_name": "ambig_qa"}, "validation": {"name": "validation", "num_bytes": 805808, "num_examples": 2002, "dataset_name": "ambig_qa"}}, "download_checksums": {"https://nlp.cs.washington.edu/ambigqa/data/ambignq_light.zip": {"num_bytes": 1061383, "checksum": "3f5dada69dec05cef1533a64945cd7bafde1aa94b0cdd6fa9a22f881206220db"}, "https://nlp.cs.washington.edu/ambigqa/data/ambignq.zip": {"num_bytes": 18639517, "checksum": "e85cec5909f076c6f584322c7f05cae44dcacaec93758c110a26fcceaa8da0ce"}}, "download_size": 19700900, "post_processing_size": null, "dataset_size": 3545540, "size_in_bytes": 23246440}, "full": {"description": "AmbigNQ, a dataset covering 14,042 questions from NQ-open, an existing open-domain QA benchmark. We find that over half of the questions in NQ-open are ambiguous. The types of ambiguity are diverse and sometimes subtle, many of which are only apparent after examining evidence provided by a very large text corpus. AMBIGNQ, a dataset with\n14,042 annotations on NQ-OPEN questions containing diverse types of ambiguity.\nWe provide two distributions of our new dataset AmbigNQ: a full version with all annotation metadata and a light version with only inputs and outputs.\n", "citation": "@inproceedings{ min2020ambigqa,\n title={ {A}mbig{QA}: Answering Ambiguous Open-domain Questions },\n author={ Min, Sewon and Michael, Julian and Hajishirzi, Hannaneh and Zettlemoyer, Luke },\n booktitle={ EMNLP },\n year={2020}\n}\n", "homepage": "https://nlp.cs.washington.edu/ambigqa/", "license": "CC BY-SA 3.0", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "annotations": {"feature": {"type": {"dtype": "string", "id": null, "_type": "Value"}, "answer": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "qaPairs": {"feature": {"question": {"dtype": "string", "id": null, "_type": "Value"}, "answer": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}}, "length": -1, "id": null, "_type": "Sequence"}}, "length": -1, "id": null, "_type": "Sequence"}, "viewed_doc_titles": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "used_queries": {"feature": {"query": {"dtype": "string", "id": null, "_type": "Value"}, "results": {"feature": {"title": {"dtype": "string", "id": null, "_type": "Value"}, "snippet": {"dtype": "string", "id": null, "_type": "Value"}}, "length": -1, "id": null, "_type": "Sequence"}}, "length": -1, "id": null, "_type": "Sequence"}, "nq_answer": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "nq_doc_title": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "builder_name": "ambig_qa", "config_name": "full", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 43538733, "num_examples": 10036, "dataset_name": "ambig_qa"}, "validation": {"name": "validation", "num_bytes": 15383368, "num_examples": 2002, "dataset_name": "ambig_qa"}}, "download_checksums": {"https://nlp.cs.washington.edu/ambigqa/data/ambignq_light.zip": {"num_bytes": 1061383, "checksum": "3f5dada69dec05cef1533a64945cd7bafde1aa94b0cdd6fa9a22f881206220db"}, "https://nlp.cs.washington.edu/ambigqa/data/ambignq.zip": {"num_bytes": 18639517, "checksum": "e85cec5909f076c6f584322c7f05cae44dcacaec93758c110a26fcceaa8da0ce"}}, "download_size": 19700900, "post_processing_size": null, "dataset_size": 58922101, "size_in_bytes": 78623001}}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "light": {
3
+ "description": "AmbigNQ, a dataset covering 14,042 questions from NQ-open, an existing open-domain QA benchmark. We find that over half of the questions in NQ-open are ambiguous. The types of ambiguity are diverse and sometimes subtle, many of which are only apparent after examining evidence provided by a very large text corpus. AMBIGNQ, a dataset with\n14,042 annotations on NQ-OPEN questions containing diverse types of ambiguity.\nWe provide two distributions of our new dataset AmbigNQ: a full version with all annotation metadata and a light version with only inputs and outputs.\n",
4
+ "citation": "@inproceedings{ min2020ambigqa,\n title={ {A}mbig{QA}: Answering Ambiguous Open-domain Questions },\n author={ Min, Sewon and Michael, Julian and Hajishirzi, Hannaneh and Zettlemoyer, Luke },\n booktitle={ EMNLP },\n year={2020}\n}\n",
5
+ "homepage": "https://nlp.cs.washington.edu/ambigqa/",
6
+ "license": "CC BY-SA 3.0",
7
+ "features": {
8
+ "id": {
9
+ "dtype": "string",
10
+ "id": null,
11
+ "_type": "Value"
12
+ },
13
+ "question": {
14
+ "dtype": "string",
15
+ "id": null,
16
+ "_type": "Value"
17
+ },
18
+ "annotations": {
19
+ "feature": {
20
+ "type": {
21
+ "dtype": "string",
22
+ "id": null,
23
+ "_type": "Value"
24
+ },
25
+ "answer": {
26
+ "feature": {
27
+ "dtype": "string",
28
+ "id": null,
29
+ "_type": "Value"
30
+ },
31
+ "length": -1,
32
+ "id": null,
33
+ "_type": "Sequence"
34
+ },
35
+ "qaPairs": {
36
+ "feature": {
37
+ "question": {
38
+ "dtype": "string",
39
+ "id": null,
40
+ "_type": "Value"
41
+ },
42
+ "answer": {
43
+ "feature": {
44
+ "dtype": "string",
45
+ "id": null,
46
+ "_type": "Value"
47
+ },
48
+ "length": -1,
49
+ "id": null,
50
+ "_type": "Sequence"
51
+ }
52
+ },
53
+ "length": -1,
54
+ "id": null,
55
+ "_type": "Sequence"
56
+ }
57
+ },
58
+ "length": -1,
59
+ "id": null,
60
+ "_type": "Sequence"
61
+ }
62
+ },
63
+ "post_processed": null,
64
+ "supervised_keys": null,
65
+ "builder_name": "ambig_qa",
66
+ "config_name": "light",
67
+ "version": {
68
+ "version_str": "1.0.0",
69
+ "description": null,
70
+ "major": 1,
71
+ "minor": 0,
72
+ "patch": 0
73
+ },
74
+ "splits": {
75
+ "train": {
76
+ "name": "train",
77
+ "num_bytes": 2739732,
78
+ "num_examples": 10036,
79
+ "dataset_name": "ambig_qa"
80
+ },
81
+ "validation": {
82
+ "name": "validation",
83
+ "num_bytes": 805808,
84
+ "num_examples": 2002,
85
+ "dataset_name": "ambig_qa"
86
+ }
87
+ },
88
+ "download_checksums": {
89
+ "https://nlp.cs.washington.edu/ambigqa/data/ambignq_light.zip": {
90
+ "num_bytes": 1061383,
91
+ "checksum": "3f5dada69dec05cef1533a64945cd7bafde1aa94b0cdd6fa9a22f881206220db"
92
+ },
93
+ "https://nlp.cs.washington.edu/ambigqa/data/ambignq.zip": {
94
+ "num_bytes": 18639517,
95
+ "checksum": "e85cec5909f076c6f584322c7f05cae44dcacaec93758c110a26fcceaa8da0ce"
96
+ }
97
+ },
98
+ "download_size": 19700900,
99
+ "post_processing_size": null,
100
+ "dataset_size": 3545540,
101
+ "size_in_bytes": 23246440
102
+ },
103
+ "full": {
104
+ "description": "AmbigNQ, a dataset covering 14,042 questions from NQ-open, an existing open-domain QA benchmark. We find that over half of the questions in NQ-open are ambiguous. The types of ambiguity are diverse and sometimes subtle, many of which are only apparent after examining evidence provided by a very large text corpus. AMBIGNQ, a dataset with\n14,042 annotations on NQ-OPEN questions containing diverse types of ambiguity.\nWe provide two distributions of our new dataset AmbigNQ: a full version with all annotation metadata and a light version with only inputs and outputs.\n",
105
+ "citation": "@inproceedings{ min2020ambigqa,\n title={ {A}mbig{QA}: Answering Ambiguous Open-domain Questions },\n author={ Min, Sewon and Michael, Julian and Hajishirzi, Hannaneh and Zettlemoyer, Luke },\n booktitle={ EMNLP },\n year={2020}\n}\n",
106
+ "homepage": "https://nlp.cs.washington.edu/ambigqa/",
107
+ "license": "CC BY-SA 3.0",
108
+ "features": {
109
+ "id": {
110
+ "dtype": "string",
111
+ "_type": "Value"
112
+ },
113
+ "question": {
114
+ "dtype": "string",
115
+ "_type": "Value"
116
+ },
117
+ "annotations": {
118
+ "feature": {
119
+ "type": {
120
+ "dtype": "string",
121
+ "_type": "Value"
122
+ },
123
+ "answer": {
124
+ "feature": {
125
+ "dtype": "string",
126
+ "_type": "Value"
127
+ },
128
+ "_type": "Sequence"
129
+ },
130
+ "qaPairs": {
131
+ "feature": {
132
+ "question": {
133
+ "dtype": "string",
134
+ "_type": "Value"
135
+ },
136
+ "answer": {
137
+ "feature": {
138
+ "dtype": "string",
139
+ "_type": "Value"
140
+ },
141
+ "_type": "Sequence"
142
+ }
143
+ },
144
+ "_type": "Sequence"
145
+ }
146
+ },
147
+ "_type": "Sequence"
148
+ },
149
+ "viewed_doc_titles": {
150
+ "feature": {
151
+ "dtype": "string",
152
+ "_type": "Value"
153
+ },
154
+ "_type": "Sequence"
155
+ },
156
+ "used_queries": {
157
+ "feature": {
158
+ "query": {
159
+ "dtype": "string",
160
+ "_type": "Value"
161
+ },
162
+ "results": {
163
+ "feature": {
164
+ "title": {
165
+ "dtype": "string",
166
+ "_type": "Value"
167
+ },
168
+ "snippet": {
169
+ "dtype": "string",
170
+ "_type": "Value"
171
+ }
172
+ },
173
+ "_type": "Sequence"
174
+ }
175
+ },
176
+ "_type": "Sequence"
177
+ },
178
+ "nq_answer": {
179
+ "feature": {
180
+ "dtype": "string",
181
+ "_type": "Value"
182
+ },
183
+ "_type": "Sequence"
184
+ },
185
+ "nq_doc_title": {
186
+ "dtype": "string",
187
+ "_type": "Value"
188
+ }
189
+ },
190
+ "builder_name": "parquet",
191
+ "dataset_name": "ambig_qa",
192
+ "config_name": "full",
193
+ "version": {
194
+ "version_str": "1.0.0",
195
+ "major": 1,
196
+ "minor": 0,
197
+ "patch": 0
198
+ },
199
+ "splits": {
200
+ "train": {
201
+ "name": "train",
202
+ "num_bytes": 43538533,
203
+ "num_examples": 10036,
204
+ "dataset_name": null
205
+ },
206
+ "validation": {
207
+ "name": "validation",
208
+ "num_bytes": 15383268,
209
+ "num_examples": 2002,
210
+ "dataset_name": null
211
+ }
212
+ },
213
+ "download_size": 30674462,
214
+ "dataset_size": 58921801,
215
+ "size_in_bytes": 89596263
216
+ }
217
+ }
full/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d95567d551a8a3101b846f184320f0cbeaecf24aca7d2acf29fbb7c54e2951f0
3
+ size 23476142
full/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc99e779bbfc3049d0f42aaf90eccc055146dd4096d68950d87568215f2a001d
3
+ size 7198320