Datasets:

Languages:
English
ArXiv:
Tags:
math
proof-pile-2 / proof-pile-2.py
Jyantkumar's picture
Update proof-pile-2.py
fa712a8 verified
raw
history blame
7.61 kB
import json
import os
import zstandard as zstd
import datasets
_CITATION="""\
@article{azerbayev2023llemma,
title={Llemma: an open language model for mathematics},
author={Zhangir Azerbayev and Hailey Schoelkopf and Keiran Paster and Marco Dos Santos and Stephen McAleer and Albert Q. Jiang and Jia Deng and Stella Biderman and Sean Welleck},
eprint={xyz.xyz},
archivePrefix={arXiv}
year={2023}
}
"""
_DESCRIPTION = """\
A dataset of high quality mathematical text. """
_HOMEPAGE = "https://github.com/EleutherAI/math-lm"
# hacky workaround: listing out files here for download, because removing dataloader script entirely introduced another bug
_ARXIV_FILES = {
"train": [f"arXiv_{i:03}.jsonl.zst" for i in range(100)],
"validation": [f"arXiv_{i:03}.jsonl.zst" for i in range(100)],
"test": [f"arXiv_{i:03}.jsonl.zst" for i in range(100)],
}
_OWM_FILES = {
"train": [f"shard-{i:04}.jsonl.zst" for i in range(63)],
"validation": ["val.jsonl.zst"],
"test": ["test.jsonl.zst"],
}
_ALGSTACK_FILES = {
"train": ["agda0000.jsonl.zst", "c0000.jsonl.zst"]
+ [f"cpp{i:04}.jsonl.zst" for i in range(5)]
+ [f"fortran{i:04}.jsonl.zst" for i in range(4)]
+ ["gap0000.jsonl.zst"]
+ [f"github-MATLAB-train-{i:04}.jsonl.zst" for i in range(4)]
+ [f"github-coq-train-{i:04}.jsonl.zst" for i in range(3)]
+ ["github-isabelle-train-0000.jsonl.zst", "github-lean-train-0000.jsonl.zst"]
+ ["haskell0000.jsonl.zst", "idris0000.jsonl.zst", "isa_proofsteps.jsonl.zst"]
+ [f"julia{i:04}.jsonl.zst" for i in range(6)]
+ ["jupyter-notebook0000.jsonl.zst", "lean_proofsteps.jsonl.zst", "maple0000.jsonl.zst"]
+ [f"python{i:04}.jsonl.zst" for i in range(42)]
+ ["r0000.jsonl.zst"]
+ [f"tex{i:04}.jsonl.zst" for i in range(3)],
"validation": [
"agda-validation.jsonl.zst",
"c-validation.jsonl.zst",
"cpp-validation.jsonl.zst",
"fortran-validation.jsonl.zst",
"gap-validation.jsonl.zst",
"github-MATLAB-validation-0000.jsonl.zst",
"github-coq-validation-0000.jsonl.zst",
"github-isabelle-validation-0000.jsonl.zst",
"github-lean-validation-0000.jsonl.zst",
"haskell-validation.jsonl.zst",
"idris-validation.jsonl.zst",
"isa_proofsteps.jsonl.zst",
"julia-validation.jsonl.zst",
"jupyter-notebook-validation.jsonl.zst",
"lean_proofsteps.jsonl.zst",
"maple-validation.jsonl.zst",
"python-validation.jsonl.zst",
"r-validation.jsonl.zst",
"tex-validation.jsonl.zst",
],
"test": [
"agda-test.jsonl.zst",
"c-test.jsonl.zst",
"cpp-test.jsonl.zst",
"fortran-test.jsonl.zst",
"gap-test.jsonl.zst",
"github-MATLAB-test-0000.jsonl.zst",
"github-coq-test-0000.jsonl.zst",
"github-isabelle-test-0000.jsonl.zst",
"github-lean-test-0000.jsonl.zst",
"haskell-test.jsonl.zst",
"idris-test.jsonl.zst",
"isa_proofsteps.jsonl.zst",
"julia-test.jsonl.zst",
"jupyter-notebook-test.jsonl.zst",
"lean_proofsteps.jsonl.zst",
"maple-test.jsonl.zst",
"python-test.jsonl.zst",
"r-test.jsonl.zst",
"tex-test.jsonl.zst",
]
}
_FILES_MAPPING = {
"arxiv": _ARXIV_FILES,
"open-web-math": _OWM_FILES,
"algebraic-stack": _ALGSTACK_FILES,
}
class ProofPile2Config(datasets.BuilderConfig):
"""BuilderConfig for RedPajama sample."""
def __init__(self, *args, subsets, **kwargs):
"""BuilderConfig for ProofPile2.
Args:
**kwargs: keyword arguments forwarded to super.
"""
super(ProofPile2Config, self).__init__(**kwargs)
self.subsets = subsets
class ProofPile2(datasets.GeneratorBasedBuilder):
"""A large dataset of mathematical text."""
VERSION = datasets.Version("1.1.0")
# This is an example of a dataset with multiple configurations.
# If you don't want/need to define several sub-sets in your dataset,
# just remove the BUILDER_CONFIG_CLASS and the BUILDER_CONFIGS attributes.
# If you need to make complex sub-parts in the datasets with configurable options
# You can create your own builder configuration class to store attribute, inheriting from ProofPile2Config
# BUILDER_CONFIG_CLASS = MyBuilderConfig
# You will be able to load one or the other configurations in the following list with
# data = datasets.load_dataset('my_dataset', 'first_domain')
# data = datasets.load_dataset('my_dataset', 'second_domain')
BUILDER_CONFIGS = [
ProofPile2Config(
name='default',
subsets=['arxiv', 'open-web-math', 'algebraic-stack'],
version=VERSION,
description="All subsets"
),
ProofPile2Config(
name='arxiv',
subsets=["arxiv"],
version=VERSION,
description="ArXiv subset"
),
ProofPile2Config(
name='open-web-math',
subsets=['open-web-math'],
version=VERSION,
description="OpenWebMath"
),
ProofPile2Config(
name='algebraic-stack',
subsets=['algebraic-stack'],
version=VERSION,
description="Code subset"
),
]
def _info(self):
features = datasets.Features(
{
"text": datasets.Value("string"),
"meta": datasets.Value("string")
}
)
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=features,
homepage=_HOMEPAGE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
return [
datasets.SplitGenerator(
name=split_obj,
# These kwargs will be passed to _generate_examples
gen_kwargs={
"data_files": list(map(
dl_manager.download_and_extract,
[
f"https://huggingface.co/datasets/EleutherAI/proof-pile-2/resolve/main/{subset}/{split}/{x}"
for subset in self.config.subsets
for x in _FILES_MAPPING[subset][split]
]
))
},
)
for split, split_obj in zip(
("train", "validation", "test"),
(datasets.Split.TRAIN, datasets.Split.VALIDATION, datasets.Split.TEST)
)
]
def _generate_examples(self, data_files):
key = 0
for name in data_files:
with zstd.open(open(name, "rb"), "rt", encoding="utf-8") as f:
for x in f.readlines():
instance = json.loads(x)
if instance:
if "meta" not in instance:
instance["meta"] = dict()
yield key, {"text": instance["text"], "meta": json.dumps(instance["meta"])}
key += 1
except:
with open(name, "r", encoding="utf-8") as f:
for x in f.readlines():
instance = json.loads(x)
if instance:
if "meta" not in instance:
instance["meta"] = dict()
yield key, {"text": instance["text"], "meta": json.dumps(instance["meta"])}
key += 1