Datasets:
GEM
/

Tasks:
Other
Multilinguality:
unknown
Size Categories:
unknown
Language Creators:
unknown
Annotations Creators:
expert-created
Source Datasets:
original
License:
mathiascreutz commited on
Commit
48f737e
1 Parent(s): 201a4b4

Added comments

Browse files
Files changed (1) hide show
  1. opusparcus.py +49 -19
opusparcus.py CHANGED
@@ -106,7 +106,6 @@ class Opusparcus(datasets.GeneratorBasedBuilder):
106
  "sent2": datasets.Value("string"),
107
  "annot_score": datasets.Value("float"),
108
  "gem_id": datasets.Value("string"),
109
- #"quality": datasets.Value("uint8")
110
  }
111
  )
112
 
@@ -130,24 +129,37 @@ class Opusparcus(datasets.GeneratorBasedBuilder):
130
 
131
  def _split_generators(self, dl_manager):
132
  """Returns SplitGenerators."""
133
- # TODO: This method is tasked with downloading/extracting the data and defining the splits depending on the configuration
134
- # If several configurations are possible (listed in BUILDER_CONFIGS), the configuration selected by the user is in self.config.name
135
- # dl_manager is a datasets.download.DownloadManager that can be used to download and extract URLs
136
- # It can accept any type or nested list/dict and will give back the same structure with the url replaced with path to local files.
137
- # By default the archives will be extracted and a path to a cached folder where they are extracted is returned instead of the archive
 
 
138
 
 
139
  if self.config.quality < 70:
140
  # We need to retrieve the largest training set file
141
  # containing the full training set for the desired language
142
  _URLs["train"] = "train_{0}.60.jsonl.bz2".format(self.config.lang)
 
143
  elif self.config.quality <= 95:
144
  # We can do with a smaller version of the training set
145
  # for the desired language
146
  _URLs["train"] = "train_{0}.70.jsonl.bz2".format(self.config.lang)
147
 
148
  # Otherwise, if the desired quality is above 95, we do not
149
- # download any training data, because there is no matching data
 
 
150
 
 
 
 
 
 
 
 
151
  data_dir = dl_manager.download_and_extract(_URLs)
152
 
153
  splits = [
@@ -193,8 +205,13 @@ class Opusparcus(datasets.GeneratorBasedBuilder):
193
  ),
194
  ]
195
 
 
 
 
 
196
  if self.config.quality <= 95:
197
- # We do have training data as well
 
198
  splits.append(
199
  datasets.SplitGenerator(
200
  name=datasets.Split.TRAIN,
@@ -211,43 +228,56 @@ class Opusparcus(datasets.GeneratorBasedBuilder):
211
  return splits
212
 
213
  def _generate_examples(
214
- self, lang, quality, filepath, split # method parameters are unpacked from `gen_kwargs` as given in `_split_generators`
 
 
215
  ):
216
-
217
  """ Yields examples as (key, example) tuples. """
218
- # This method handles input defined in _split_generators to yield (key, example) tuples from the dataset.
 
219
  # The `key` is here for legacy reason (tfds) and is not important in itself.
 
220
  if split == datasets.Split.TRAIN:
 
 
 
 
 
221
  with bz2.open(filepath, "rt", encoding="utf-8") as f:
222
- # We know that this file only contains the desired language,
223
- # because for the training sets the languages are in separate
224
- # files, and only the desired language has been downloaded
225
  for id_, row in enumerate(f):
226
  data = json.loads(row)
227
  if data["quality"] < quality:
228
- # The rest of this file contains too low quality data
 
229
  break
230
  yield id_, {
231
  "lang": data["lang"],
232
  "sent1": data["sent1"],
233
  "sent2": data["sent2"],
234
- "annot_score": 0.0,
235
  "gem_id": data["gem_id"],
236
- #"quality": data["quality"],
237
  }
238
  else:
 
 
 
 
 
239
  keep_all = (split == "validation.full" or split == "test.full")
240
  with open(filepath, encoding="utf-8") as f:
241
  for id_, row in enumerate(f):
242
  data = json.loads(row)
243
- if data["lang"] == lang:
244
  if keep_all or data["annot_score"] >= 3.0:
 
 
 
 
245
  yield id_, {
246
  "lang": data["lang"],
247
  "sent1": data["sent1"],
248
  "sent2": data["sent2"],
249
  "annot_score": data["annot_score"],
250
  "gem_id": data["gem_id"],
251
- #"quality": 100,
252
  }
253
 
 
106
  "sent2": datasets.Value("string"),
107
  "annot_score": datasets.Value("float"),
108
  "gem_id": datasets.Value("string"),
 
109
  }
110
  )
111
 
 
129
 
130
  def _split_generators(self, dl_manager):
131
  """Returns SplitGenerators."""
132
+ # This method is tasked with downloading/extracting the data
133
+ # and defining the splits depending on the configuration.
134
+ # Several configurations are possible (listed in
135
+ # BUILDER_CONFIGS), and the configuration selected by the user
136
+ # is in self.config.name, which consists of two fields
137
+ # separated by a period, containing the values of
138
+ # self.config.lang and self.config.quality.
139
 
140
+ # Select which file of the training data contains the matching data:
141
  if self.config.quality < 70:
142
  # We need to retrieve the largest training set file
143
  # containing the full training set for the desired language
144
  _URLs["train"] = "train_{0}.60.jsonl.bz2".format(self.config.lang)
145
+
146
  elif self.config.quality <= 95:
147
  # We can do with a smaller version of the training set
148
  # for the desired language
149
  _URLs["train"] = "train_{0}.70.jsonl.bz2".format(self.config.lang)
150
 
151
  # Otherwise, if the desired quality is above 95, we do not
152
+ # download any training data, because there is no matching data.
153
+ # The validation and test sets are so small that we do not perform
154
+ # any filtering or optimization at this stage.
155
 
156
+ # dl_manager is a datasets.download.DownloadManager, which
157
+ # downloads and extracts the URLs
158
+ # (It can accept any type or nested list/dict and will give
159
+ # back the same structure with the url replaced with path to
160
+ # local files. By default the archives will be extracted and
161
+ # a path to a cached folder where they are extracted is
162
+ # returned instead of the archive.)
163
  data_dir = dl_manager.download_and_extract(_URLs)
164
 
165
  splits = [
 
205
  ),
206
  ]
207
 
208
+ # If the desired quality value is 100, no subset of the
209
+ # training set is good enough, and we only produce validation
210
+ # and test sets, in order to save space and time.
211
+
212
  if self.config.quality <= 95:
213
+ # In this case there is matching training data, so we produce
214
+ # a train split.
215
  splits.append(
216
  datasets.SplitGenerator(
217
  name=datasets.Split.TRAIN,
 
228
  return splits
229
 
230
  def _generate_examples(
231
+ self, lang, quality, filepath, split
232
+ # method parameters are unpacked from `gen_kwargs` as given in
233
+ # `_split_generators`
234
  ):
 
235
  """ Yields examples as (key, example) tuples. """
236
+ # This method handles input defined in _split_generators to
237
+ # yield (key, example) tuples from the dataset.
238
  # The `key` is here for legacy reason (tfds) and is not important in itself.
239
+
240
  if split == datasets.Split.TRAIN:
241
+ # Training sets are in compressed bz2 files.
242
+ # They contain a field "quality" missing from the validation and test sets.
243
+ # We also know that this file only contains the desired language,
244
+ # because for the training sets the languages are in separate
245
+ # files, and only the desired language has been downloaded.
246
  with bz2.open(filepath, "rt", encoding="utf-8") as f:
 
 
 
247
  for id_, row in enumerate(f):
248
  data = json.loads(row)
249
  if data["quality"] < quality:
250
+ # The rest of this file contains too low quality data,
251
+ # because the data is sorted best first
252
  break
253
  yield id_, {
254
  "lang": data["lang"],
255
  "sent1": data["sent1"],
256
  "sent2": data["sent2"],
257
+ "annot_score": 0.0, # means there is no annotation
258
  "gem_id": data["gem_id"],
 
259
  }
260
  else:
261
+ # The validation and test sets are in jsonl files.
262
+ # They contain the fields "lang" and "quality" that we filter on.
263
+ # If we ask for the full sets, we will keep all data entries, also
264
+ # the sentence pairs that were not considered paraphrases by the
265
+ # annotators:
266
  keep_all = (split == "validation.full" or split == "test.full")
267
  with open(filepath, encoding="utf-8") as f:
268
  for id_, row in enumerate(f):
269
  data = json.loads(row)
270
+ if data["lang"] == lang: # only keep desired language
271
  if keep_all or data["annot_score"] >= 3.0:
272
+ # for full sets keep all;
273
+ # for standard test and validation sets, keep only
274
+ # the paraphrases (annot_score >= 3.0 means "good
275
+ # or mostly good example of paraphrases")
276
  yield id_, {
277
  "lang": data["lang"],
278
  "sent1": data["sent1"],
279
  "sent2": data["sent2"],
280
  "annot_score": data["annot_score"],
281
  "gem_id": data["gem_id"],
 
282
  }
283