Datasets:
ArXiv:
License:
#!/usr/bin/python3 | |
# -*- coding: utf-8 -*- | |
from glob import glob | |
import json | |
import os | |
from pathlib import Path | |
import datasets | |
_URLS = { | |
"amazon_reviews_multi": "data/amazon_reviews_multi.jsonl", | |
"bucc2018": "data/bucc2018.jsonl", | |
"mike0307": "data/mike0307.jsonl", | |
"nbnn": "data/nbnn.jsonl", | |
"nordic_langid": "data/nordic_langid.jsonl", | |
"scandi_langid": "data/scandi_langid.jsonl", | |
"stsb_multi_mt": "data/stsb_multi_mt.jsonl", | |
"tatoeba": "data/tatoeba.jsonl", | |
"xnli": "data/xnli.jsonl", | |
} | |
_CITATION = """\ | |
@dataset{language_identification, | |
author = {Xing Tian}, | |
title = {language_identification}, | |
month = aug, | |
year = 2024, | |
publisher = {Xing Tian}, | |
version = {1.0}, | |
} | |
""" | |
LANGUAGE_MAP = { | |
"ar": "arabic", | |
"bg": "bulgarian", | |
"da": "danish", | |
"de": "german", | |
"el": "modern greek", | |
"en": "english", | |
"eo": "esperanto", | |
"es": "spanish", | |
"fo": "faroese", | |
"fr": "french", | |
"gl": "galician", | |
"hi": "hindi", | |
"is": "icelandic", | |
"it": "italian", | |
"ja": "japanese", | |
"nl": "dutch", | |
"no": "norwegian", | |
"no-b": "norwegian (bokmål)", | |
"no-n": "norwegian (nynorsk)", | |
"mr": "marathi", | |
"pl": "polish", | |
"pt": "portuguese", | |
"ru": "russian", | |
"sw": "swahili", | |
"sv": "swedish", | |
"th": "thai", | |
"tr": "turkish", | |
"ur": "urdu", | |
"vi": "vietnamese", | |
"zh": "chinese", | |
"zh-cn": "simplified chinese", | |
"zh-tw": "traditional chinese", | |
} | |
class LanguageIdentification(datasets.GeneratorBasedBuilder): | |
VERSION = datasets.Version("1.0.0") | |
BUILDER_CONFIGS = [ | |
datasets.BuilderConfig(name="amazon_reviews_multi", version=VERSION, description="amazon_reviews_multi"), | |
datasets.BuilderConfig(name="bucc2018", version=VERSION, description="bucc2018"), | |
datasets.BuilderConfig(name="mike0307", version=VERSION, description="mike0307"), | |
datasets.BuilderConfig(name="nbnn", version=VERSION, description="nbnn"), | |
datasets.BuilderConfig(name="nordic_langid", version=VERSION, description="nordic_langid"), | |
datasets.BuilderConfig(name="scandi_langid", version=VERSION, description="scandi_langid"), | |
datasets.BuilderConfig(name="stsb_multi_mt", version=VERSION, description="stsb_multi_mt"), | |
datasets.BuilderConfig(name="tatoeba", version=VERSION, description="tatoeba"), | |
datasets.BuilderConfig(name="xnli", version=VERSION, description="xnli"), | |
] | |
def _info(self): | |
features = datasets.Features( | |
{ | |
"text": datasets.Value("string"), | |
"language": datasets.Value("string"), | |
"data_source": datasets.Value("string"), | |
} | |
) | |
return datasets.DatasetInfo( | |
features=features, | |
supervised_keys=None, | |
homepage="", | |
license="", | |
citation=_CITATION, | |
) | |
def _split_generators(self, dl_manager): | |
"""Returns SplitGenerators.""" | |
url = _URLS[self.config.name] | |
dl_path = dl_manager.download(url) | |
archive_path = dl_path | |
return [ | |
datasets.SplitGenerator( | |
name=datasets.Split.TRAIN, | |
gen_kwargs={"archive_path": archive_path, "split": "train"}, | |
), | |
datasets.SplitGenerator( | |
name=datasets.Split.VALIDATION, | |
gen_kwargs={"archive_path": archive_path, "split": "validation"}, | |
), | |
datasets.SplitGenerator( | |
name=datasets.Split.TEST, | |
gen_kwargs={"archive_path": archive_path, "split": "test"}, | |
), | |
] | |
def _generate_examples(self, archive_path, split): | |
archive_path = Path(archive_path) | |
idx = 0 | |
with open(archive_path, "r", encoding="utf-8") as f: | |
for row in f: | |
sample = json.loads(row) | |
if sample["split"] != split: | |
continue | |
language = sample["language"] | |
if language not in LANGUAGE_MAP.keys(): | |
raise AssertionError(language) | |
yield idx, { | |
"text": sample["text"], | |
"language": language, | |
"data_source": sample["data_source"], | |
} | |
idx += 1 | |
if __name__ == '__main__': | |
pass | |