Datasets:
Multilinguality:
monolingual
Size Categories:
1M<n<10M
Language Creators:
found
Annotations Creators:
no-annotation
Source Datasets:
original
ArXiv:
Tags:
License:
import os | |
import datasets | |
_CITATION = """\ | |
@article{cruz2020investigating, | |
title={Investigating the True Performance of Transformers in Low-Resource Languages: A Case Study in Automatic Corpus Creation}, | |
author={Jan Christian Blaise Cruz and Jose Kristian Resabal and James Lin and Dan John Velasco and Charibeth Cheng}, | |
journal={arXiv preprint arXiv:2010.11574}, | |
year={2020} | |
} | |
""" | |
_DESCRIPTION = """\ | |
Large-scale dataset of Filipino news articles. Sourced for the NewsPH-NLI Project (Cruz et al., 2020). | |
""" | |
_URL = "https://github.com/jcblaisecruz02/Filipino-Text-Benchmarks" | |
_LICENSE = "GPL-3.0" | |
_DATA_URL = "https://s3.us-east-2.amazonaws.com/blaisecruz.com/datasets/newsph" | |
class NewsphConfig(datasets.BuilderConfig): | |
def __init__(self, data_url, **kwargs): | |
super(NewsphConfig, self).__init__( | |
version=datasets.Version( | |
"1.0.0", | |
), | |
**kwargs, | |
) | |
self.data_url = data_url | |
class Newsph(datasets.GeneratorBasedBuilder): | |
VERSION = datasets.Version("1.0.0") | |
BUILDER_CONFIGS = [ | |
NewsphConfig( | |
name="newsph", | |
data_url=_DATA_URL + "/" + "newsph.zip", | |
description=_DESCRIPTION, | |
), | |
] | |
BUILDER_CONFIG_CLASS = NewsphConfig | |
def _info(self): | |
return datasets.DatasetInfo( | |
# This is the description that will appear on the datasets page. | |
description=_DESCRIPTION, | |
features=datasets.Features({"text": datasets.Value("string")}), | |
supervised_keys=None, | |
# Homepage of the dataset for documentation | |
homepage=_URL, | |
citation=_CITATION, | |
license=_LICENSE, | |
) | |
def _split_generators(self, dl_manager): | |
"""Returns SplitGenerators.""" | |
data_file = dl_manager.download_and_extract(self.config.data_url) | |
data_dir = os.path.join(data_file, "newsph") | |
return [ | |
datasets.SplitGenerator( | |
name=datasets.Split.TRAIN, | |
gen_kwargs={"data_file": os.path.join(data_dir, "train.txt"), "split": "train"}, | |
), | |
] | |
def _generate_examples(self, data_file, split): | |
"""Yields examples.""" | |
with open(data_file, encoding="utf-8") as f: | |
for idx, row in enumerate(f): | |
if row.strip(): | |
yield idx, {"text": row} | |
else: | |
yield idx, {"text": ""} | |