Datasets:
Tasks:
Translation
Multilinguality:
translation
Language Creators:
found
Annotations Creators:
found
Source Datasets:
original
ArXiv:
Tags:
License:
# coding=utf-8 | |
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor. | |
# | |
# Licensed under the Apache License, Version 2.0 (the "License"); | |
# you may not use this file except in compliance with the License. | |
# You may obtain a copy of the License at | |
# | |
# http://www.apache.org/licenses/LICENSE-2.0 | |
# | |
# Unless required by applicable law or agreed to in writing, software | |
# distributed under the License is distributed on an "AS IS" BASIS, | |
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | |
# See the License for the specific language governing permissions and | |
# limitations under the License. | |
"""Bianet: A parallel news corpus in Turkish, Kurdish and English""" | |
import os | |
import datasets | |
_CITATION = """\ | |
@InProceedings{ATAMAN18.6, | |
author = {Duygu Ataman}, | |
title = {Bianet: A Parallel News Corpus in Turkish, Kurdish and English}, | |
booktitle = {Proceedings of the Eleventh International Conference on Language Resources and Evaluation (LREC 2018)}, | |
year = {2018}, | |
month = {may}, | |
date = {7-12}, | |
location = {Miyazaki, Japan}, | |
editor = {Jinhua Du and Mihael Arcan and Qun Liu and Hitoshi Isahara}, | |
publisher = {European Language Resources Association (ELRA)}, | |
address = {Paris, France}, | |
isbn = {979-10-95546-15-3}, | |
language = {english} | |
}""" | |
_HOMEPAGE = "http://opus.nlpl.eu/Bianet.php" | |
_LICENSE = "CC-BY-SA-4.0" | |
_VALID_LANGUAGE_PAIRS = { | |
("en", "ku"): "http://opus.nlpl.eu/download.php?f=Bianet/v1/moses/en-ku.txt.zip", | |
("en", "tr"): "http://opus.nlpl.eu/download.php?f=Bianet/v1/moses/en-tr.txt.zip", | |
("ku", "tr"): "http://opus.nlpl.eu/download.php?f=Bianet/v1/moses/ku-tr.txt.zip", | |
} | |
_VERSION = "1.0.0" | |
_DESCRIPTION = """\ | |
A parallel news corpus in Turkish, Kurdish and English. | |
Bianet collects 3,214 Turkish articles with their sentence-aligned Kurdish or English translations from the Bianet online newspaper. | |
3 languages, 3 bitexts | |
total number of files: 6 | |
total number of tokens: 2.25M | |
total number of sentence fragments: 0.14M | |
""" | |
_BASE_NAME = "Bianet.{}-{}.{}" | |
class BianetConfig(datasets.BuilderConfig): | |
"""BuilderConfig for Bianet: A parallel news corpus in Turkish, Kurdish and English""" | |
def __init__(self, language_pair=(None, None), **kwargs): | |
"""BuilderConfig for Bianet: A parallel news corpus in Turkish, Kurdish and English. | |
The first language in `language_pair` should consist of two strings joined by | |
an underscore (e.g. "en-tr"). | |
Args: | |
language_pair: pair of languages that will be used for translation. | |
**kwargs: keyword arguments forwarded to super. | |
""" | |
name = "%s_to_%s" % (language_pair[0], language_pair[1]) | |
description = ("Translation dataset from %s to %s or %s to %s.") % ( | |
language_pair[0], | |
language_pair[1], | |
language_pair[1], | |
language_pair[0], | |
) | |
super(BianetConfig, self).__init__( | |
name=name, description=description, version=datasets.Version(_VERSION, ""), **kwargs | |
) | |
# Validate language pair. | |
assert language_pair in _VALID_LANGUAGE_PAIRS, ( | |
"Config language pair (%s, " "%s) not supported" | |
) % language_pair | |
self.language_pair = language_pair | |
class Bianet(datasets.GeneratorBasedBuilder): | |
BUILDER_CONFIGS = [ | |
BianetConfig( | |
language_pair=pair, | |
) | |
for pair in _VALID_LANGUAGE_PAIRS.keys() | |
] | |
BUILDER_CONFIG_CLASS = BianetConfig | |
def _info(self): | |
return datasets.DatasetInfo( | |
description=_DESCRIPTION, | |
features=datasets.Features( | |
{ | |
"id": datasets.Value("string"), | |
"translation": datasets.Translation(languages=tuple(self.config.language_pair)), | |
}, | |
), | |
supervised_keys=None, | |
homepage=_HOMEPAGE, | |
citation=_CITATION, | |
) | |
def _split_generators(self, dl_manager): | |
download_url = _VALID_LANGUAGE_PAIRS.get(tuple(self.config.language_pair)) | |
path = dl_manager.download_and_extract(download_url) | |
return [ | |
datasets.SplitGenerator( | |
name=datasets.Split.TRAIN, | |
gen_kwargs={"datapath": path}, | |
) | |
] | |
def _generate_examples(self, datapath): | |
lang1, lang2 = self.config.language_pair | |
lang1_file = _BASE_NAME.format(lang1, lang2, lang1) | |
lang2_file = _BASE_NAME.format(lang1, lang2, lang2) | |
lang1_path = os.path.join(datapath, lang1_file) | |
lang2_path = os.path.join(datapath, lang2_file) | |
with open(lang1_path, encoding="utf-8") as f1, open(lang2_path, encoding="utf-8") as f2: | |
for sentence_counter, (x, y) in enumerate(zip(f1, f2)): | |
x = x.strip() | |
y = y.strip() | |
result = ( | |
sentence_counter, | |
{ | |
"id": str(sentence_counter), | |
"translation": {lang1: x, lang2: y}, | |
}, | |
) | |
yield result | |