|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""The Open WebText Corpus""" |
|
|
|
import os |
|
import re |
|
import tarfile |
|
|
|
import datasets |
|
|
|
|
|
_CITATION = """\ |
|
@misc{Gokaslan2019OpenWeb, |
|
title={OpenWebText Corpus}, |
|
author={Aaron Gokaslan*, Vanya Cohen*, Ellie Pavlick, Stefanie Tellex}, |
|
howpublished{\\url{http://Skylion007.github.io/OpenWebTextCorpus}}, |
|
year={2019} |
|
} |
|
""" |
|
|
|
_DESCRIPTION = """\ |
|
An open-source replication of the WebText dataset from OpenAI. |
|
""" |
|
|
|
_N_DATA_FILES = 21 |
|
_DATA_FILES = ["subsets/urlsf_subset{:02d}.tar".format(i) for i in range(_N_DATA_FILES)] |
|
|
|
|
|
def _iter_tar(f): |
|
stream = tarfile.open(fileobj=f, mode="r|*") |
|
for tarinfo in stream: |
|
file_path = tarinfo.name |
|
if not tarinfo.isreg(): |
|
continue |
|
if file_path is None: |
|
continue |
|
if os.path.basename(file_path).startswith((".", "__")): |
|
|
|
continue |
|
file_obj = stream.extractfile(tarinfo) |
|
yield file_path, file_obj |
|
stream.members = [] |
|
del stream |
|
|
|
|
|
class Openwebtext(datasets.GeneratorBasedBuilder): |
|
"""The Open WebText dataset.""" |
|
|
|
BUILDER_CONFIGS = [ |
|
datasets.BuilderConfig( |
|
name="plain_text", |
|
description="Plain text", |
|
version=datasets.Version("1.0.0"), |
|
) |
|
] |
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=datasets.Features({"text": datasets.Value("string")}), |
|
homepage="https://skylion007.github.io/OpenWebTextCorpus/", |
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
archives = dl_manager.download(_DATA_FILES) |
|
return [ |
|
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={ |
|
"archive_iterators": [ |
|
dl_manager.iter_archive(archive) for archive in archives |
|
] |
|
}), |
|
] |
|
|
|
def _generate_examples(self, archive_iterators): |
|
"""Yields examples.""" |
|
for archive_iterator in archive_iterators: |
|
for xz_filepath, xz_f in archive_iterator: |
|
if not xz_filepath.endswith(".xz"): |
|
continue |
|
for txt_filepath, txt_f in _iter_tar(xz_f): |
|
if not txt_filepath.endswith(".txt"): |
|
continue |
|
idx = f"{xz_filepath}/{txt_filepath}" |
|
yield idx, {"text": re.sub("\n\n\n+", "\n\n", txt_f.read().decode("utf-8")).strip()} |
|
|