|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""GitHub Code dataset.""" |
|
|
|
import os |
|
|
|
import pyarrow as pa |
|
import pyarrow.parquet as pq |
|
|
|
import datasets |
|
from huggingface_hub import HfApi, HfFolder |
|
from datasets.data_files import DataFilesDict |
|
|
|
_REPO_NAME = "lvwerra/github-code" |
|
|
|
_LANG_TO_EXTENSION = { |
|
"Assembly": [".asm"], |
|
"Batchfile": [".bat", ".cmd"], |
|
"C": [".c", ".h"], |
|
"C#": [".cs"], |
|
"C++": [".cpp", ".hpp", ".c++", ".h++", ".cc", ".hh", ".C", ".H"], |
|
"CMake": [".cmake"], |
|
"CSS": [".css"], |
|
"Dockerfile": [".dockerfile", "Dockerfile"], |
|
"FORTRAN": ['.f90', '.f', '.f03', '.f08', '.f77', '.f95', '.for', '.fpp'], |
|
"GO": [".go"], |
|
"Haskell": [".hs"], |
|
"HTML":[".html"], |
|
"Java": [".java"], |
|
"JavaScript": [".js"], |
|
"Julia": [".jl"], |
|
"Lua": [".lua"], |
|
"Makefile": ["Makefile"], |
|
"Markdown": [".md", ".markdown"], |
|
"PHP": [".php", ".php3", ".php4", ".php5", ".phps", ".phpt"], |
|
"Perl": [".pl", ".pm", ".pod", ".perl"], |
|
"PowerShell": ['.ps1', '.psd1', '.psm1'], |
|
"Python": [".py"], |
|
"Ruby": [".rb"], |
|
"Rust": [".rs"], |
|
"SQL": [".sql"], |
|
"Scala": [".scala"], |
|
"Shell": [".sh", ".bash", ".command", ".zsh"], |
|
"TypeScript": [".ts", ".tsx"], |
|
"TeX": [".tex"], |
|
"Visual Basic": [".vb"] |
|
} |
|
|
|
_LICENSES = ['mit', |
|
'apache-2.0', |
|
'gpl-3.0', |
|
'gpl-2.0', |
|
'bsd-3-clause', |
|
'agpl-3.0', |
|
'lgpl-3.0', |
|
'lgpl-2.1', |
|
'bsd-2-clause', |
|
'cc0-1.0', |
|
'epl-1.0', |
|
'mpl-2.0', |
|
'unlicense', |
|
'isc', |
|
'artistic-2.0'] |
|
|
|
_DESCRIPTION = """\ |
|
The GitHub Code dataest consists of 115M code files from GitHub in 32 programming \ |
|
languages with 60 extensions totalling in 1TB of text data. The dataset was created \ |
|
from the GitHub dataset on BiqQuery. |
|
""" |
|
|
|
_HOMEPAGE = "https://cloud.google.com/blog/topics/public-datasets/github-on-bigquery-analyze-all-the-open-source-code/" |
|
|
|
|
|
_EXTENSION_TO_LANG = {} |
|
for lang in _LANG_TO_EXTENSION: |
|
for extension in _LANG_TO_EXTENSION[lang]: |
|
_EXTENSION_TO_LANG[extension] = lang |
|
|
|
|
|
|
|
_LANG_CONFIGS = ["all"] + list(_LANG_TO_EXTENSION.keys()) |
|
_LICENSE_CONFIGS = ["all"] + _LICENSES |
|
|
|
class GithubCodeConfig(datasets.BuilderConfig): |
|
"""BuilderConfig for the GitHub Code dataset.""" |
|
|
|
def __init__(self, *args, languages=["all"], licenses=["all"], **kwargs): |
|
"""BuilderConfig for the GitHub Code dataset. |
|
|
|
Args: |
|
languages (:obj:`List[str]`): List of languages to load. |
|
licenses (:obj:`List[str]`): List of licenses to load. |
|
**kwargs: keyword arguments forwarded to super. |
|
""" |
|
super().__init__( |
|
*args, |
|
name="+".join(languages)+"-"+"+".join(licenses), |
|
**kwargs, |
|
) |
|
|
|
languages = set(languages) |
|
licenses = set(licenses) |
|
|
|
assert all([language in _LANG_CONFIGS for language in languages]), f"Language not in {_LANG_CONFIGS}." |
|
assert all([license in _LICENSE_CONFIGS for license in licenses]), f"License not in {_LICENSE_CONFIGS}." |
|
|
|
if "all" in languages: |
|
assert len(languages)==1, "Passed 'all' together with other languages." |
|
self.filter_languages = False |
|
else: |
|
self.filter_languages = True |
|
|
|
if "all" in licenses: |
|
assert len(licenses)==1, "Passed 'all' together with other licenses." |
|
self.filter_licenses = False |
|
else: |
|
self.filter_licenses = True |
|
|
|
self.languages = set(languages) |
|
self.licenses = set(licenses) |
|
|
|
|
|
|
|
class GithubCode(datasets.GeneratorBasedBuilder): |
|
"""GitHub Code dataset.""" |
|
|
|
VERSION = datasets.Version("1.0.0") |
|
|
|
BUILDER_CONFIG_CLASS = GithubCodeConfig |
|
BUILDER_CONFIGS = [GithubCodeConfig(languages=[lang], licenses=[license]) for lang in _LANG_CONFIGS |
|
for license in _LICENSE_CONFIGS] |
|
DEFAULT_CONFIG_NAME = "all-all" |
|
|
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=datasets.Features({"code": datasets.Value("string"), |
|
"repo_name": datasets.Value("string"), |
|
"path": datasets.Value("string"), |
|
"language": datasets.Value("string"), |
|
"license": datasets.Value("string"), |
|
"size": datasets.Value("int32")}), |
|
supervised_keys=None, |
|
homepage=_HOMEPAGE, |
|
license="Multiple: see the 'license' field of each sample.", |
|
|
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
|
|
hfh_dataset_info = HfApi(datasets.config.HF_ENDPOINT).dataset_info( |
|
_REPO_NAME, |
|
timeout=100.0, |
|
) |
|
|
|
patterns = datasets.data_files.get_patterns_in_dataset_repository(hfh_dataset_info) |
|
data_files = datasets.data_files.DataFilesDict.from_hf_repo( |
|
patterns, |
|
dataset_info=hfh_dataset_info, |
|
) |
|
|
|
files = dl_manager.download_and_extract(data_files["train"]) |
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
gen_kwargs={ |
|
"files": files, |
|
}, |
|
), |
|
] |
|
|
|
def _generate_examples(self, files): |
|
key = 0 |
|
for file_idx, file in enumerate(files): |
|
with open(file, "rb") as f: |
|
parquet_file = pq.ParquetFile(f) |
|
for batch_idx, record_batch in enumerate(parquet_file.iter_batches(batch_size=10_000)): |
|
pa_table = pa.Table.from_batches([record_batch]) |
|
for row_index in range(pa_table.num_rows): |
|
row = pa_table.slice(row_index, 1).to_pydict() |
|
|
|
lang = lang_from_name(row['path'][0]) |
|
license = row["license"][0] |
|
|
|
if self.config.filter_languages and not lang in self.config.languages: |
|
continue |
|
if self.config.filter_licenses and not license in self.config.licenses: |
|
continue |
|
|
|
yield key, {"code": row['content'][0], |
|
"repo_name": row['repo_name'][0], |
|
"path": row['path'][0], |
|
"license": license, |
|
"language": lang, |
|
"size": int(row['size'][0])} |
|
key += 1 |
|
|
|
|
|
def lang_from_name(name): |
|
for extension in _EXTENSION_TO_LANG: |
|
if name.endswith(extension): |
|
return _EXTENSION_TO_LANG[extension] |