albertvillanova HF staff commited on
Commit
42d29b5
1 Parent(s): 06f0711

Convert dataset to Parquet (#2)

Browse files

- Convert dataset to Parquet (50b3fed0bd821e5c8ccacd10743721c67c70788c)
- Add QDMR-high-level data files (81631d5f56e9b32c08e1eb77e40de7ae6c728c07)
- Add QDMR-high-level-lexicon data files (c78ffd0009b7c111eabd5a2a8eef75f9671f3132)
- Add QDMR-lexicon data files (22d685e8dd62a43c48efa3d17896540fc7ae7da7)
- Add logical-forms data files (0441e53ab8827b30684f0ac56e20d7f072c3e456)
- Delete loading script (8e45a818e0a58112331d499d55325325f7ddbeea)
- Delete legacy dataset_infos.json (0d4defb8a57b79d3f69442a86bf65143fc069dab)

QDMR-high-level-lexicon/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bc44d768f77e269d6825ebc8e803c9d47be2d28fe0d402e36864e45579c13d7
3
+ size 759218
QDMR-high-level-lexicon/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42d7a05e22b078ac2f148474dc153599346f957680bfdf3ecbb879589becdef0
3
+ size 4159209
QDMR-high-level-lexicon/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:145fe7cf7c047e8d719464bd78b06147b1a6d63d4c252ac288ae6b2d01f8017e
3
+ size 745497
QDMR-high-level/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edf00b230e5f1478bcc55f4e10d751c98d330e732debf5f7885630d3f1ffdbf8
3
+ size 262748
QDMR-high-level/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea809a3aaeb31982a4a64ef5c07aeab30ef4cd878fcd199a4f10408e429e4f5f
3
+ size 2412936
QDMR-high-level/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d21ca330865ec917233df3cd6e7f2f3ecf43a7bf8ccfe377635426b27909adb
3
+ size 437503
QDMR-lexicon/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c6be50e96716795cc51daa1aff25c07c4d3550358f3143912d36e4ad4ec7108
3
+ size 1459038
QDMR-lexicon/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dc6cce0e2e5c1c865aaca22296b2a278262e0a7373a35150d1eace8554948ec
3
+ size 7950250
QDMR-lexicon/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ceda93b7a2849be9afa6b237f7b097c482272f59b02401f90e5275603c7e889a
3
+ size 1408978
QDMR/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f692a11f225ee86d0645db7fec5c530d373bc317a69b4eac5320382150162c55
3
+ size 373459
QDMR/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e27bb7b0de5c2623e1fbc6f2f1c80d53bdc18746bd786ea850e826fe2df175b0
3
+ size 4063484
QDMR/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df1a02f6f4b631487c9048009de85fc7e3cb9758f1ff1bf5fc88ae105230c836
3
+ size 738565
README.md CHANGED
@@ -20,7 +20,7 @@ task_ids:
20
  paperswithcode_id: break
21
  pretty_name: BREAK
22
  dataset_info:
23
- - config_name: QDMR-high-level
24
  features:
25
  - name: question_id
26
  dtype: string
@@ -33,36 +33,18 @@ dataset_info:
33
  - name: split
34
  dtype: string
35
  splits:
36
- - name: test
37
- num_bytes: 482339
38
- num_examples: 3195
39
  - name: train
40
- num_bytes: 5148086
41
- num_examples: 17503
42
  - name: validation
43
- num_bytes: 914780
44
- num_examples: 3130
45
- download_size: 15971078
46
- dataset_size: 6545205
47
- - config_name: QDMR-high-level-lexicon
48
- features:
49
- - name: source
50
- dtype: string
51
- - name: allowed_tokens
52
- dtype: string
53
- splits:
54
  - name: test
55
- num_bytes: 4240755
56
- num_examples: 3195
57
- - name: train
58
- num_bytes: 23234518
59
- num_examples: 17503
60
- - name: validation
61
- num_bytes: 4158679
62
- num_examples: 3130
63
- download_size: 15971078
64
- dataset_size: 31633952
65
- - config_name: QDMR
66
  features:
67
  - name: question_id
68
  dtype: string
@@ -75,17 +57,35 @@ dataset_info:
75
  - name: split
76
  dtype: string
77
  splits:
 
 
 
 
 
 
78
  - name: test
79
- num_bytes: 900632
80
- num_examples: 8069
 
 
 
 
 
 
 
 
 
81
  - name: train
82
- num_bytes: 12790466
83
- num_examples: 44321
84
  - name: validation
85
- num_bytes: 2237472
86
- num_examples: 7760
87
- download_size: 15971078
88
- dataset_size: 15928570
 
 
 
89
  - config_name: QDMR-lexicon
90
  features:
91
  - name: source
@@ -93,17 +93,17 @@ dataset_info:
93
  - name: allowed_tokens
94
  dtype: string
95
  splits:
96
- - name: test
97
- num_bytes: 10331822
98
- num_examples: 8069
99
  - name: train
100
- num_bytes: 56913064
101
  num_examples: 44321
102
  - name: validation
103
- num_bytes: 9936933
104
  num_examples: 7760
105
- download_size: 15971078
106
- dataset_size: 77181819
 
 
 
107
  - config_name: logical-forms
108
  features:
109
  - name: question_id
@@ -119,17 +119,58 @@ dataset_info:
119
  - name: program
120
  dtype: string
121
  splits:
122
- - name: test
123
- num_bytes: 927038
124
- num_examples: 8006
125
  - name: train
126
- num_bytes: 19821676
127
  num_examples: 44098
128
  - name: validation
129
- num_bytes: 3504893
130
  num_examples: 7719
131
- download_size: 15971078
132
- dataset_size: 24253607
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
133
  ---
134
 
135
  # Dataset Card for "break_data"
 
20
  paperswithcode_id: break
21
  pretty_name: BREAK
22
  dataset_info:
23
+ - config_name: QDMR
24
  features:
25
  - name: question_id
26
  dtype: string
 
33
  - name: split
34
  dtype: string
35
  splits:
 
 
 
36
  - name: train
37
+ num_bytes: 12757200
38
+ num_examples: 44321
39
  - name: validation
40
+ num_bytes: 2231632
41
+ num_examples: 7760
 
 
 
 
 
 
 
 
 
42
  - name: test
43
+ num_bytes: 894558
44
+ num_examples: 8069
45
+ download_size: 5175508
46
+ dataset_size: 15883390
47
+ - config_name: QDMR-high-level
 
 
 
 
 
 
48
  features:
49
  - name: question_id
50
  dtype: string
 
57
  - name: split
58
  dtype: string
59
  splits:
60
+ - name: train
61
+ num_bytes: 5134938
62
+ num_examples: 17503
63
+ - name: validation
64
+ num_bytes: 912408
65
+ num_examples: 3130
66
  - name: test
67
+ num_bytes: 479919
68
+ num_examples: 3195
69
+ download_size: 3113187
70
+ dataset_size: 6527265
71
+ - config_name: QDMR-high-level-lexicon
72
+ features:
73
+ - name: source
74
+ dtype: string
75
+ - name: allowed_tokens
76
+ dtype: string
77
+ splits:
78
  - name: train
79
+ num_bytes: 23227946
80
+ num_examples: 17503
81
  - name: validation
82
+ num_bytes: 4157495
83
+ num_examples: 3130
84
+ - name: test
85
+ num_bytes: 4239547
86
+ num_examples: 3195
87
+ download_size: 5663924
88
+ dataset_size: 31624988
89
  - config_name: QDMR-lexicon
90
  features:
91
  - name: source
 
93
  - name: allowed_tokens
94
  dtype: string
95
  splits:
 
 
 
96
  - name: train
97
+ num_bytes: 56896433
98
  num_examples: 44321
99
  - name: validation
100
+ num_bytes: 9934015
101
  num_examples: 7760
102
+ - name: test
103
+ num_bytes: 10328787
104
+ num_examples: 8069
105
+ download_size: 10818266
106
+ dataset_size: 77159235
107
  - config_name: logical-forms
108
  features:
109
  - name: question_id
 
119
  - name: program
120
  dtype: string
121
  splits:
 
 
 
122
  - name: train
123
+ num_bytes: 19783061
124
  num_examples: 44098
125
  - name: validation
126
+ num_bytes: 3498114
127
  num_examples: 7719
128
+ - name: test
129
+ num_bytes: 920007
130
+ num_examples: 8006
131
+ download_size: 7572815
132
+ dataset_size: 24201182
133
+ configs:
134
+ - config_name: QDMR
135
+ data_files:
136
+ - split: train
137
+ path: QDMR/train-*
138
+ - split: validation
139
+ path: QDMR/validation-*
140
+ - split: test
141
+ path: QDMR/test-*
142
+ - config_name: QDMR-high-level
143
+ data_files:
144
+ - split: train
145
+ path: QDMR-high-level/train-*
146
+ - split: validation
147
+ path: QDMR-high-level/validation-*
148
+ - split: test
149
+ path: QDMR-high-level/test-*
150
+ - config_name: QDMR-high-level-lexicon
151
+ data_files:
152
+ - split: train
153
+ path: QDMR-high-level-lexicon/train-*
154
+ - split: validation
155
+ path: QDMR-high-level-lexicon/validation-*
156
+ - split: test
157
+ path: QDMR-high-level-lexicon/test-*
158
+ - config_name: QDMR-lexicon
159
+ data_files:
160
+ - split: train
161
+ path: QDMR-lexicon/train-*
162
+ - split: validation
163
+ path: QDMR-lexicon/validation-*
164
+ - split: test
165
+ path: QDMR-lexicon/test-*
166
+ - config_name: logical-forms
167
+ data_files:
168
+ - split: train
169
+ path: logical-forms/train-*
170
+ - split: validation
171
+ path: logical-forms/validation-*
172
+ - split: test
173
+ path: logical-forms/test-*
174
  ---
175
 
176
  # Dataset Card for "break_data"
break_data.py DELETED
@@ -1,261 +0,0 @@
1
- """TODO(break_data): Add a description here."""
2
-
3
-
4
- import csv
5
- import json
6
- import os
7
- import textwrap
8
-
9
- import datasets
10
-
11
-
12
- # TODO(break): BibTeX citation
13
- _CITATION = """\
14
- @article{Wolfson2020Break,
15
- title={Break It Down: A Question Understanding Benchmark},
16
- author={Wolfson, Tomer and Geva, Mor and Gupta, Ankit and Gardner, Matt and Goldberg, Yoav and Deutch, Daniel and Berant, Jonathan},
17
- journal={Transactions of the Association for Computational Linguistics},
18
- year={2020},
19
- }
20
- """
21
-
22
- # TODO(break):
23
- _DESCRIPTION = """\
24
- Break is a human annotated dataset of natural language questions and their Question Decomposition Meaning Representations
25
- (QDMRs). Break consists of 83,978 examples sampled from 10 question answering datasets over text, images and databases.
26
- This repository contains the Break dataset along with information on the exact data format.
27
- """
28
- _URL = "https://github.com/allenai/Break/raw/master/break_dataset/Break-dataset.zip"
29
-
30
-
31
- class BreakDataConfig(datasets.BuilderConfig):
32
-
33
- """BuilderConfig for Break"""
34
-
35
- def __init__(self, text_features, lexicon_tokens, **kwargs):
36
- """
37
-
38
- Args:
39
- text_features: `dict[string, string]`, map from the name of the feature
40
- dict for each text field to the name of the column in the tsv file
41
- lexicon_tokens: to define if we want to load the lexicon_tokens files or not
42
- **kwargs: keyword arguments forwarded to super.
43
- """
44
- super(BreakDataConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs)
45
- self.text_features = text_features
46
- self.lexicon_tokens = lexicon_tokens
47
-
48
-
49
- class BreakData(datasets.GeneratorBasedBuilder):
50
- """TODO(break_data): Short description of my dataset."""
51
-
52
- # TODO(break_data): Set up version.
53
- VERSION = datasets.Version("0.1.0")
54
- BUILDER_CONFIGS = [
55
- BreakDataConfig(
56
- name="QDMR-high-level",
57
- description=textwrap.dedent(
58
- """
59
- Contains questions annotated with the high-level variant of QDMR. These decomposition are exclusive to Reading
60
- Comprehension tasks (Section 2). lexicon_tokens files are also provided."""
61
- ),
62
- text_features={
63
- "question_id": "question_id",
64
- "question_text": "question_text",
65
- "decomposition": "decomposition",
66
- "operators": "operators",
67
- "split": "split",
68
- },
69
- lexicon_tokens=False,
70
- ),
71
- BreakDataConfig(
72
- name="QDMR-high-level-lexicon",
73
- description=textwrap.dedent(
74
- """
75
- Contains questions annotated with the high-level variant of QDMR. These decomposition are exclusive to Reading
76
- Comprehension tasks (Section 2). lexicon_tokens files are also provided."""
77
- ),
78
- text_features={
79
- "source": "source",
80
- "allowed_tokens": "allowed_tokens",
81
- },
82
- lexicon_tokens=True,
83
- ),
84
- BreakDataConfig(
85
- name="QDMR",
86
- description=textwrap.dedent(
87
- """
88
- Contains questions over text, images and databases annotated with their Question Decomposition Meaning
89
- Representation. In addition to the train, dev and (hidden) test sets we provide lexicon_tokens files. For
90
- each question, the lexicon file contains the set of valid tokens that could potentially appear in its
91
- decomposition """
92
- ),
93
- text_features={
94
- "question_id": "question_id",
95
- "question_text": "question_text",
96
- "decomposition": "decomposition",
97
- "operators": "operators",
98
- "split": "split",
99
- },
100
- lexicon_tokens=False,
101
- ),
102
- BreakDataConfig(
103
- name="QDMR-lexicon",
104
- description=textwrap.dedent(
105
- """
106
- Contains questions over text, images and databases annotated with their Question Decomposition Meaning
107
- Representation. In addition to the train, dev and (hidden) test sets we provide lexicon_tokens files. For
108
- each question, the lexicon file contains the set of valid tokens that could potentially appear in its
109
- decomposition """
110
- ),
111
- text_features={
112
- "source": "source",
113
- "allowed_tokens": "allowed_tokens",
114
- },
115
- lexicon_tokens=True,
116
- ),
117
- BreakDataConfig(
118
- name="logical-forms",
119
- description=textwrap.dedent(
120
- """
121
- Contains questions and QDMRs annotated with full logical-forms of QDMR operators + arguments. Full logical-forms
122
- were inferred by the annotation-consistency algorithm described in """
123
- ),
124
- lexicon_tokens=False,
125
- text_features={
126
- "question_id": "question_id",
127
- "question_text": "question_text",
128
- "decomposition": "decomposition",
129
- "operators": "operators",
130
- "split": "split",
131
- "program": "program",
132
- },
133
- ),
134
- ]
135
-
136
- def _info(self):
137
- # TODO(break_data): Specifies the datasets.DatasetInfo object
138
- features = {text_feature: datasets.Value("string") for text_feature in self.config.text_features.keys()}
139
- return datasets.DatasetInfo(
140
- # This is the description that will appear on the datasets page.
141
- description=_DESCRIPTION,
142
- # datasets.features.FeatureConnectors
143
- features=datasets.Features(
144
- features
145
- # These are the features of your dataset like images, labels ...
146
- ),
147
- # If there's a common (input, target) tuple from the features,
148
- # specify them here. They'll be used if as_supervised=True in
149
- # builder.as_dataset.
150
- supervised_keys=None,
151
- # Homepage of the dataset for documentation
152
- homepage="https://github.com/allenai/Break",
153
- citation=_CITATION,
154
- )
155
- # if
156
-
157
- def _split_generators(self, dl_manager):
158
- """Returns SplitGenerators."""
159
- # TODO(break_data): Downloads the data and defines the splits
160
- # dl_manager is a datasets.download.DownloadManager that can be used to
161
- # download and extract URLs
162
- dl_dir = dl_manager.download_and_extract(_URL)
163
- data_dir = os.path.join(dl_dir, "Break-dataset")
164
- qdmr_high_level = os.path.join(data_dir, "QDMR-high-level")
165
- qdmr = os.path.join(data_dir, "QDMR")
166
- logical = os.path.join(data_dir, "logical-forms")
167
- if self.config.name == "QDMR" or self.config.name == "QDMR-lexicon":
168
- return [
169
- datasets.SplitGenerator(
170
- name=datasets.Split.TRAIN,
171
- # These kwargs will be passed to _generate_examples
172
- gen_kwargs={
173
- "filepath": os.path.join(qdmr, "train.csv")
174
- if not self.config.lexicon_tokens
175
- else os.path.join(qdmr, "train_lexicon_tokens.json")
176
- },
177
- ),
178
- datasets.SplitGenerator(
179
- name=datasets.Split.VALIDATION,
180
- # These kwargs will be passed to _generate_examples
181
- gen_kwargs={
182
- "filepath": os.path.join(qdmr, "dev.csv")
183
- if not self.config.lexicon_tokens
184
- else os.path.join(qdmr, "dev_lexicon_tokens.json")
185
- },
186
- ),
187
- datasets.SplitGenerator(
188
- name=datasets.Split.TEST,
189
- # These kwargs will be passed to _generate_examples
190
- gen_kwargs={
191
- "filepath": os.path.join(qdmr, "test.csv")
192
- if not self.config.lexicon_tokens
193
- else os.path.join(qdmr, "test_lexicon_tokens.json")
194
- },
195
- ),
196
- ]
197
- elif self.config.name == "QDMR-high-level" or self.config.name == "QDMR-high-level-lexicon":
198
- return [
199
- datasets.SplitGenerator(
200
- name=datasets.Split.TRAIN,
201
- # These kwargs will be passed to _generate_examples
202
- gen_kwargs={
203
- "filepath": os.path.join(qdmr_high_level, "train.csv")
204
- if not self.config.lexicon_tokens
205
- else os.path.join(qdmr_high_level, "train_lexicon_tokens.json")
206
- },
207
- ),
208
- datasets.SplitGenerator(
209
- name=datasets.Split.VALIDATION,
210
- # These kwargs will be passed to _generate_examples
211
- gen_kwargs={
212
- "filepath": os.path.join(qdmr_high_level, "dev.csv")
213
- if not self.config.lexicon_tokens
214
- else os.path.join(qdmr_high_level, "dev_lexicon_tokens.json")
215
- },
216
- ),
217
- datasets.SplitGenerator(
218
- name=datasets.Split.TEST,
219
- # These kwargs will be passed to _generate_examples
220
- gen_kwargs={
221
- "filepath": os.path.join(qdmr_high_level, "test.csv")
222
- if not self.config.lexicon_tokens
223
- else os.path.join(qdmr_high_level, "test_lexicon_tokens.json")
224
- },
225
- ),
226
- ]
227
- elif self.config.name == "logical-forms":
228
- return [
229
- datasets.SplitGenerator(
230
- name=datasets.Split.TRAIN,
231
- # These kwargs will be passed to _generate_examples
232
- gen_kwargs={"filepath": os.path.join(logical, "train.csv")},
233
- ),
234
- datasets.SplitGenerator(
235
- name=datasets.Split.VALIDATION,
236
- # These kwargs will be passed to _generate_examples
237
- gen_kwargs={"filepath": os.path.join(logical, "dev.csv")},
238
- ),
239
- datasets.SplitGenerator(
240
- name=datasets.Split.TEST,
241
- # These kwargs will be passed to _generate_examples
242
- gen_kwargs={"filepath": os.path.join(logical, "test.csv")},
243
- ),
244
- ]
245
-
246
- def _generate_examples(self, filepath):
247
- """Yields examples."""
248
- # TODO(break_data): Yields (key, example) tuples from the dataset
249
- with open(filepath, encoding="utf-8") as f:
250
- if (
251
- self.config.name == "QDMR-high-level"
252
- or self.config.name == "QDMR"
253
- or self.config.name == "logical-forms"
254
- ):
255
- data = csv.DictReader(f)
256
- for id_, row in enumerate(data):
257
- yield id_, row
258
- elif self.config.name == "QDMR-high-level-lexicon" or self.config.name == "QDMR-lexicon":
259
- for id_, row in enumerate(f):
260
- data = json.loads(row)
261
- yield id_, data
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dataset_infos.json DELETED
@@ -1 +0,0 @@
1
- {"QDMR-high-level": {"description": "Break is a human annotated dataset of natural language questions and their Question Decomposition Meaning Representations\n(QDMRs). Break consists of 83,978 examples sampled from 10 question answering datasets over text, images and databases. \nThis repository contains the Break dataset along with information on the exact data format.\n", "citation": "@article{Wolfson2020Break,\n title={Break It Down: A Question Understanding Benchmark},\n author={Wolfson, Tomer and Geva, Mor and Gupta, Ankit and Gardner, Matt and Goldberg, Yoav and Deutch, Daniel and Berant, Jonathan},\n journal={Transactions of the Association for Computational Linguistics},\n year={2020},\n}\n", "homepage": "https://github.com/allenai/Break", "license": "", "features": {"question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_text": {"dtype": "string", "id": null, "_type": "Value"}, "decomposition": {"dtype": "string", "id": null, "_type": "Value"}, "operators": {"dtype": "string", "id": null, "_type": "Value"}, "split": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "break_data", "config_name": "QDMR-high-level", "version": {"version_str": "1.0.0", "description": "", "datasets_version_to_prepare": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 482339, "num_examples": 3195, "dataset_name": "break_data"}, "train": {"name": "train", "num_bytes": 5148086, "num_examples": 17503, "dataset_name": "break_data"}, "validation": {"name": "validation", "num_bytes": 914780, "num_examples": 3130, "dataset_name": "break_data"}}, "download_checksums": {"https://github.com/allenai/Break/raw/master/break_dataset/Break-dataset.zip": {"num_bytes": 15971078, "checksum": "37efea4fa1b7774d077ff0452e5e199cecba8216c12da76781010f189d1cf259"}}, "download_size": 15971078, "dataset_size": 6545205, "size_in_bytes": 22516283}, "QDMR-high-level-lexicon": {"description": "Break is a human annotated dataset of natural language questions and their Question Decomposition Meaning Representations\n(QDMRs). Break consists of 83,978 examples sampled from 10 question answering datasets over text, images and databases. \nThis repository contains the Break dataset along with information on the exact data format.\n", "citation": "@article{Wolfson2020Break,\n title={Break It Down: A Question Understanding Benchmark},\n author={Wolfson, Tomer and Geva, Mor and Gupta, Ankit and Gardner, Matt and Goldberg, Yoav and Deutch, Daniel and Berant, Jonathan},\n journal={Transactions of the Association for Computational Linguistics},\n year={2020},\n}\n", "homepage": "https://github.com/allenai/Break", "license": "", "features": {"source": {"dtype": "string", "id": null, "_type": "Value"}, "allowed_tokens": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "break_data", "config_name": "QDMR-high-level-lexicon", "version": {"version_str": "1.0.0", "description": "", "datasets_version_to_prepare": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 4240755, "num_examples": 3195, "dataset_name": "break_data"}, "train": {"name": "train", "num_bytes": 23234518, "num_examples": 17503, "dataset_name": "break_data"}, "validation": {"name": "validation", "num_bytes": 4158679, "num_examples": 3130, "dataset_name": "break_data"}}, "download_checksums": {"https://github.com/allenai/Break/raw/master/break_dataset/Break-dataset.zip": {"num_bytes": 15971078, "checksum": "37efea4fa1b7774d077ff0452e5e199cecba8216c12da76781010f189d1cf259"}}, "download_size": 15971078, "dataset_size": 31633952, "size_in_bytes": 47605030}, "QDMR": {"description": "Break is a human annotated dataset of natural language questions and their Question Decomposition Meaning Representations\n(QDMRs). Break consists of 83,978 examples sampled from 10 question answering datasets over text, images and databases. \nThis repository contains the Break dataset along with information on the exact data format.\n", "citation": "@article{Wolfson2020Break,\n title={Break It Down: A Question Understanding Benchmark},\n author={Wolfson, Tomer and Geva, Mor and Gupta, Ankit and Gardner, Matt and Goldberg, Yoav and Deutch, Daniel and Berant, Jonathan},\n journal={Transactions of the Association for Computational Linguistics},\n year={2020},\n}\n", "homepage": "https://github.com/allenai/Break", "license": "", "features": {"question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_text": {"dtype": "string", "id": null, "_type": "Value"}, "decomposition": {"dtype": "string", "id": null, "_type": "Value"}, "operators": {"dtype": "string", "id": null, "_type": "Value"}, "split": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "break_data", "config_name": "QDMR", "version": {"version_str": "1.0.0", "description": "", "datasets_version_to_prepare": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 900632, "num_examples": 8069, "dataset_name": "break_data"}, "train": {"name": "train", "num_bytes": 12790466, "num_examples": 44321, "dataset_name": "break_data"}, "validation": {"name": "validation", "num_bytes": 2237472, "num_examples": 7760, "dataset_name": "break_data"}}, "download_checksums": {"https://github.com/allenai/Break/raw/master/break_dataset/Break-dataset.zip": {"num_bytes": 15971078, "checksum": "37efea4fa1b7774d077ff0452e5e199cecba8216c12da76781010f189d1cf259"}}, "download_size": 15971078, "dataset_size": 15928570, "size_in_bytes": 31899648}, "QDMR-lexicon": {"description": "Break is a human annotated dataset of natural language questions and their Question Decomposition Meaning Representations\n(QDMRs). Break consists of 83,978 examples sampled from 10 question answering datasets over text, images and databases. \nThis repository contains the Break dataset along with information on the exact data format.\n", "citation": "@article{Wolfson2020Break,\n title={Break It Down: A Question Understanding Benchmark},\n author={Wolfson, Tomer and Geva, Mor and Gupta, Ankit and Gardner, Matt and Goldberg, Yoav and Deutch, Daniel and Berant, Jonathan},\n journal={Transactions of the Association for Computational Linguistics},\n year={2020},\n}\n", "homepage": "https://github.com/allenai/Break", "license": "", "features": {"source": {"dtype": "string", "id": null, "_type": "Value"}, "allowed_tokens": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "break_data", "config_name": "QDMR-lexicon", "version": {"version_str": "1.0.0", "description": "", "datasets_version_to_prepare": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 10331822, "num_examples": 8069, "dataset_name": "break_data"}, "train": {"name": "train", "num_bytes": 56913064, "num_examples": 44321, "dataset_name": "break_data"}, "validation": {"name": "validation", "num_bytes": 9936933, "num_examples": 7760, "dataset_name": "break_data"}}, "download_checksums": {"https://github.com/allenai/Break/raw/master/break_dataset/Break-dataset.zip": {"num_bytes": 15971078, "checksum": "37efea4fa1b7774d077ff0452e5e199cecba8216c12da76781010f189d1cf259"}}, "download_size": 15971078, "dataset_size": 77181819, "size_in_bytes": 93152897}, "logical-forms": {"description": "Break is a human annotated dataset of natural language questions and their Question Decomposition Meaning Representations\n(QDMRs). Break consists of 83,978 examples sampled from 10 question answering datasets over text, images and databases. \nThis repository contains the Break dataset along with information on the exact data format.\n", "citation": "@article{Wolfson2020Break,\n title={Break It Down: A Question Understanding Benchmark},\n author={Wolfson, Tomer and Geva, Mor and Gupta, Ankit and Gardner, Matt and Goldberg, Yoav and Deutch, Daniel and Berant, Jonathan},\n journal={Transactions of the Association for Computational Linguistics},\n year={2020},\n}\n", "homepage": "https://github.com/allenai/Break", "license": "", "features": {"question_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_text": {"dtype": "string", "id": null, "_type": "Value"}, "decomposition": {"dtype": "string", "id": null, "_type": "Value"}, "operators": {"dtype": "string", "id": null, "_type": "Value"}, "split": {"dtype": "string", "id": null, "_type": "Value"}, "program": {"dtype": "string", "id": null, "_type": "Value"}}, "supervised_keys": null, "builder_name": "break_data", "config_name": "logical-forms", "version": {"version_str": "1.0.0", "description": "", "datasets_version_to_prepare": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 927038, "num_examples": 8006, "dataset_name": "break_data"}, "train": {"name": "train", "num_bytes": 19821676, "num_examples": 44098, "dataset_name": "break_data"}, "validation": {"name": "validation", "num_bytes": 3504893, "num_examples": 7719, "dataset_name": "break_data"}}, "download_checksums": {"https://github.com/allenai/Break/raw/master/break_dataset/Break-dataset.zip": {"num_bytes": 15971078, "checksum": "37efea4fa1b7774d077ff0452e5e199cecba8216c12da76781010f189d1cf259"}}, "download_size": 15971078, "dataset_size": 24253607, "size_in_bytes": 40224685}}
 
 
logical-forms/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:688eba7877766326ced484cbd5e41bcec9774f12627809886ee788dedaef5fa8
3
+ size 373456
logical-forms/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:913ddd7dc4606e532fda5b777e7b385497b6387c0d2da325d0a117d6f59738b3
3
+ size 6088626
logical-forms/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb322a97453430cd46a9d1c14c4078107f4c34ea484a952467bacfe688acc0c1
3
+ size 1110733