bayelemabaga / bayelemabaga.py
s7d11's picture
Upload bayelemabaga.py
3608d44
raw
history blame
4.27 kB
""" RobotsMaliAI: Bayelemaba """
import datasets
_CITATION = """\
@misc{bayelemabagamldataset2022
title={Machine Learning Dataset Development for Manding Languages},
author={
Valentin Vydrin and
Christopher Homan and
Michael Leventhal and
Allashera Auguste Tapo and
Marco Zampieri and
Jean-Jacques Meric and
Kirill Maslinsky and
Andrij Rovenchak and
Sebastien Diarra
},
howpublished = {url{https://github.com/robotsmali-ai/datasets}},
year={2022}
}
"""
_DESCRIPTION = """\
The Bayelemabaga dataset is a collection of 44160 aligned machine translation ready Bambara-French lines,
originating from Corpus Bambara de Reference. The dataset is constitued of text extracted from 231 source files,
varing from periodicals, books, short stories, blog posts, part of the Bible and the Quran.
"""
_URL = {
"parallel": "https://robotsmali-ai.github.io/datasets/bayelemabaga.tar.gz"
}
_LanguagePairs = [
"bam-fr", "fr-bam"]
class BayelemabagaConfig(datasets.BuilderConfig):
""" BuilderConfig for Bayelemabaga """
def __init__(self, language_pair, **kwargs) -> None:
"""
Args:
language_pair: language pair, you want to load
**kwargs: -> Super()
"""
super().__init__(**kwargs)
self.language_pair = language_pair
class Bayelemabaga(datasets.GeneratorBasedBuilder):
""" Bi-Lingual Bam, Fr text made for Machine Translation """
VERSION = datasets.Version("1.0.0")
BUILDER_CONFIG_CLASS = BayelemabagaConfig
BUILDER_CONFIGS = [
BayelemabagaConfig(name="bam-fr", description=_DESCRIPTION, language_pair="bam-fr"),
BayelemabagaConfig(name="fr-bam", description=_DESCRIPTION, language_pair="fr-bam")
]
def _info(self):
src_tag, tgt_tag = self.config.language_pair.split("-")
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=datasets.Features({"translation": datasets.features.Translation(languages=(src_tag, tgt_tag))}),
supervised_keys=(src_tag, tgt_tag),
homepage="https://robotsmali-ai.github.io/datasets",
citation=_CITATION
)
def _split_generators(self, dl_manager):
lang_pair = self.config.language_pair
src_tag, tgt_tag = lang_pair.split("-")
archive = dl_manager.download(_URL["parallel"])
train_dir = "bayelemabaga/train"
valid_dir = "bayelemabaga/valid"
test_dir = "bayelemabaga/test"
train = datasets.SplitGenerator(
name=datasets.Split.TRAIN,
gen_kwargs = {
"filepath": f"{train_dir}/train.{src_tag}",
"labelpath": f"{train_dir}/train.{tgt_tag}",
"files": dl_manager.iter_archive(archive)
}
)
valid = datasets.SplitGenerator(
name=datasets.Split.VALIDATION,
gen_kwargs = {
"filepath": f"{valid_dir}/dev.{src_tag}",
"labelpath": f"{valid_dir}/dev.{tgt_tag}",
"files": dl_manager.iter_archive(archive)
}
)
test = datasets.SplitGenerator(
name=datasets.Split.TEST,
gen_kwargs = {
"filepath": f"{test_dir}/test.{src_tag}",
"labelpath": f"{test_dir}/test.{tgt_tag}",
"files": dl_manager.iter_archive(archive)
}
)
output = []
output.append(train)
output.append(valid)
output.append(test)
return output
def _generate_examples(self, filepath, labelpath, files):
""" Yield examples """
src_tag, tgt_tag = self.config.language_pair.split("-")
src, tgt = None, None
for path, f in files:
if(path == filepath):
src = f.read().decode("utf-8").split("\n")[:-1]
elif(path == labelpath):
tgt = f.read().decode("utf-8").split("\n")[:-1]
if(src is not None and tgt is not None):
for idx, (s,t) in enumerate(zip(src, tgt)):
yield idx, {"translation": {src_tag: s, tgt_tag: t}}
break