europarl / europarl_bilingual.py
albertvillanova's picture
Fix bug in language pairs (#7)
e2dce49 verified
# coding=utf-8
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import itertools
import os
import xml.etree.ElementTree as ET
import datasets
# Find for instance the citation on arxiv or on the dataset repo/website
_CITATION = """\
@inproceedings{koehn-2005-europarl,
title = "{E}uroparl: A Parallel Corpus for Statistical Machine Translation",
author = "Koehn, Philipp",
booktitle = "Proceedings of Machine Translation Summit X: Papers",
month = sep # " 13-15",
year = "2005",
address = "Phuket, Thailand",
url = "https://aclanthology.org/2005.mtsummit-papers.11",
pages = "79--86",
}
@inproceedings{tiedemann-2012-parallel,
title = "Parallel Data, Tools and Interfaces in {OPUS}",
author = {Tiedemann, J{\\"o}rg},
editor = "Calzolari, Nicoletta and
Choukri, Khalid and
Declerck, Thierry and
Do{\\u{g}}an, Mehmet U{\\u{g}}ur and
Maegaard, Bente and
Mariani, Joseph and
Moreno, Asuncion and
Odijk, Jan and
Piperidis, Stelios",
booktitle = "Proceedings of the Eighth International Conference on Language Resources and Evaluation ({LREC}'12)",
month = may,
year = "2012",
address = "Istanbul, Turkey",
publisher = "European Language Resources Association (ELRA)",
url = "http://www.lrec-conf.org/proceedings/lrec2012/pdf/463_Paper.pdf",
pages = "2214--2218",
}"""
# You can copy an official description
_DESCRIPTION = """\
A parallel corpus extracted from the European Parliament web site by Philipp Koehn (University of Edinburgh). The main intended use is to aid statistical machine translation research.
"""
# Add a link to an official homepage for the dataset here
_HOMEPAGE = "https://opus.nlpl.eu/Europarl/corpus/version/Europarl"
# Add the licence for the dataset here if you can find it
_LICENSE = """\
The data set comes with the same license
as the original sources.
Please, check the information about the source
that is given on
https://opus.nlpl.eu/Europarl/corpus/version/Europarl
"""
# The HuggingFace dataset library don't host the datasets but only point to the original files
# This can be an arbitrary nested dict/list of URLs (see below in `_split_generators` method)
LANGUAGES = [
"bg",
"cs",
"da",
"de",
"el",
"en",
"es",
"et",
"fi",
"fr",
"hu",
"it",
"lt",
"lv",
"nl",
"pl",
"pt",
"ro",
"sk",
"sl",
"sv",
]
LANGUAGE_PAIRS = list(itertools.combinations(LANGUAGES, 2))
_VERSION = "8.0.0"
_BASE_URL_DATASET = "https://object.pouta.csc.fi/OPUS-Europarl/v8/raw/{}.zip"
_BASE_URL_RELATIONS = "https://object.pouta.csc.fi/OPUS-Europarl/v8/xml/{}-{}.xml.gz"
class EuroparlBilingualConfig(datasets.BuilderConfig):
"""Slightly custom config to require source and target languages."""
def __init__(self, *args, lang1=None, lang2=None, **kwargs):
super().__init__(
*args,
name=f"{lang1}-{lang2}",
**kwargs,
)
self.lang1 = lang1
self.lang2 = lang2
def _lang_pair(self):
return (self.lang1, self.lang2)
def _is_valid(self):
return self._lang_pair() in LANGUAGE_PAIRS
class EuroparlBilingual(datasets.GeneratorBasedBuilder):
"""Europarl contains aligned sentences in multiple west language pairs."""
VERSION = datasets.Version(_VERSION)
BUILDER_CONFIG_CLASS = EuroparlBilingualConfig
BUILDER_CONFIGS = [
EuroparlBilingualConfig(lang1=lang1, lang2=lang2, version=datasets.Version(_VERSION))
for lang1, lang2 in LANGUAGE_PAIRS
]
def _info(self):
"""This method specifies the datasets.DatasetInfo object which contains informations and typings for the dataset."""
features = datasets.Features(
{
"translation": datasets.Translation(languages=(self.config.lang1, self.config.lang2)),
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
supervised_keys=None,
homepage=_HOMEPAGE,
license=_LICENSE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
"""Returns SplitGenerators."""
if not self.config._is_valid():
raise ValueError(
f"{self.config._lang_pair()} is not a supported language pair. Choose among: {LANGUAGE_PAIRS}"
)
# download data files
path_datafile_1 = dl_manager.download_and_extract(_BASE_URL_DATASET.format(self.config.lang1))
path_datafile_2 = dl_manager.download_and_extract(_BASE_URL_DATASET.format(self.config.lang2))
# download relations file
path_relation_file = dl_manager.download_and_extract(
_BASE_URL_RELATIONS.format(self.config.lang1, self.config.lang2)
)
return [
datasets.SplitGenerator(
name=datasets.Split.TRAIN,
# These kwargs will be passed to _generate_examples
gen_kwargs={
"path_datafiles": (path_datafile_1, path_datafile_2),
"path_relation_file": path_relation_file,
},
)
]
@staticmethod
def _parse_xml_datafile(filepath):
"""
Parse and return a Dict[sentence_id, text] representing data with the following structure:
"""
document = ET.parse(filepath).getroot()
return {tag.attrib["id"]: tag.text for tag in document.iter("s")}
def _generate_examples(self, path_datafiles, path_relation_file):
"""Yields examples.
In parenthesis the useful attributes
Lang files XML
- document
- CHAPTER ('ID')
- P ('id')
- s ('id')
Relation file XML
- cesAlign
- linkGrp ('fromDoc', 'toDoc')
- link ('xtargets': '1;1')
"""
# my counter
_id = 0
relations_root = ET.parse(path_relation_file).getroot()
for linkGroup in relations_root:
# retrieve files and remove .gz extension because 'datasets' library already decompress them
from_doc_dict = EuroparlBilingual._parse_xml_datafile(
os.path.splitext(os.path.join(path_datafiles[0], "Europarl", "raw", linkGroup.attrib["fromDoc"]))[0]
)
to_doc_dict = EuroparlBilingual._parse_xml_datafile(
os.path.splitext(os.path.join(path_datafiles[1], "Europarl", "raw", linkGroup.attrib["toDoc"]))[0]
)
for link in linkGroup:
from_sentence_ids, to_sentence_ids = link.attrib["xtargets"].split(";")
from_sentence_ids = [i for i in from_sentence_ids.split(" ") if i]
to_sentence_ids = [i for i in to_sentence_ids.split(" ") if i]
if not len(from_sentence_ids) or not len(to_sentence_ids):
continue
# in rare cases, there is not entry for some key pairs
sentence_lang1 = " ".join(from_doc_dict[i] for i in from_sentence_ids if i in from_doc_dict)
sentence_lang2 = " ".join(to_doc_dict[i] for i in to_sentence_ids if i in to_doc_dict)
yield _id, {"translation": {self.config.lang1: sentence_lang1, self.config.lang2: sentence_lang2}}
_id += 1