"""LR-Sum summarization dataset""" import json import os import datasets _CITATION = """\ @inproceedings{palen-michel-lignos-2023-lr, title = "{LR}-Sum: Summarization for Less-Resourced Languages", author = "Palen-Michel, Chester and Lignos, Constantine", booktitle = "Findings of the Association for Computational Linguistics: ACL 2023", month = jul, year = "2023", address = "Toronto, Canada", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/2023.findings-acl.427", doi = "10.18653/v1/2023.findings-acl.427", pages = "6829--6844", abstract = "We introduce LR-Sum, a new permissively-licensed dataset created with the goal of enabling further research in automatic summarization for less-resourced languages.LR-Sum contains human-written summaries for 40 languages, many of which are less-resourced. We describe our process for extracting and filtering the dataset from the Multilingual Open Text corpus (Palen-Michel et al., 2022).The source data is public domain newswire collected from from Voice of America websites, and LR-Sum is released under a Creative Commons license (CC BY 4.0), making it one of the most openly-licensed multilingual summarization datasets. We describe abstractive and extractive summarization experiments to establish baselines and discuss the limitations of this dataset.", } """ _DESCRIPTION = """\ We introduce LR-Sum, a new permissively-licensed dataset created with the goal of enabling further research in automatic summarization for less-resourced languages. LR-Sum contains human-written summaries for 40 languages, many of which are less-resourced. We describe our process for extracting and filtering the dataset from the Multilingual Open Text corpus (Palen-Michel et al., 2022). The source data is public domain newswire collected from from Voice of America websites, and LR-Sum is released under a Creative Commons license (CC BY 4.0), making it one of the most openly-licensed multilingual summarization datasets. We describe abstractive and extractive summarization experiments to establish baselines and discuss the limitations of this dataset. """ _HOMEPAGE = "https://github.com/bltlab" _LICENSE = "Creative Commons Attribution 4.0 International (CC-BY 4.0)" _URL = "https://huggingface.co/datasets/bltlab/lr-sum/resolve/main/data/{}.tar.bz2" _LANGUAGES = [ "amh", "aze", "ben", "bod", "bos", "ckb", "cmn_t", "cmn_s", "ell", "eng", "fas", "fra", "hat", "hau", "hye", "ind", "kat", "khm", "kin", "kor", "kmr", "lao", "mkd", "mya", "nde", "por", "prs", "pus", "rus", "sna", "som", "spa", "sqi", "srp", "swh", "tha", "tir", "tur", "ukr", "urd", "uzb", "vie", ] class Lrsum(datasets.GeneratorBasedBuilder): VERSION = datasets.Version("1.0.0") BUILDER_CONFIGS = [ datasets.BuilderConfig( name="{}".format(lang), version=datasets.Version("1.0.0") ) for lang in _LANGUAGES ] def _info(self): return datasets.DatasetInfo( description=_DESCRIPTION, features=datasets.Features( { "id": datasets.Value("string"), "url": datasets.Value("string"), "title": datasets.Value("string"), "summary": datasets.Value("string"), "text": datasets.Value("string"), } ), supervised_keys=None, homepage=_HOMEPAGE, citation=_CITATION, license=_LICENSE, version=self.VERSION, ) def _split_generators(self, dl_manager): """Returns SplitGenerators.""" lang = str(self.config.name) url = _URL.format(lang) data_dir = dl_manager.download_and_extract(url) ret = [ datasets.SplitGenerator( name=datasets.Split.TEST, gen_kwargs={ "filepath": os.path.join(data_dir, lang, lang + "_test.jsonl"), }, ) ] if os.path.exists(os.path.join(data_dir, lang, lang + "_train.jsonl")): ret.append(datasets.SplitGenerator( name=datasets.Split.TRAIN, gen_kwargs={ "filepath": os.path.join(data_dir, lang, lang + "_train.jsonl"), }, ) ) if os.path.exists(os.path.join(data_dir, lang, lang + "_val.jsonl")): ret.append( datasets.SplitGenerator( name=datasets.Split.VALIDATION, gen_kwargs={ "filepath": os.path.join(data_dir, lang, lang + "_val.jsonl"), }, ) ) return ret def _generate_examples(self, filepath): """Yields examples as (key, example) tuples.""" with open(filepath, encoding="utf-8") as f: for idx_, row in enumerate(f): data = json.loads(row) yield idx_, { "id": data["id_"], "url": data["url"], "title": data["title"], "summary": data["summary"], "text": data["text"], }