Datasets:
Languages:
code
Multilinguality:
monolingual
Size Categories:
100K<n<1M
Language Creators:
found
Annotations Creators:
machine-generated
Source Datasets:
original
License:
pretty_name: PyAst | |
annotations_creators: | |
- machine-generated | |
language_creators: | |
- found | |
language: | |
- code | |
license: | |
- bsd-2-clause | |
- mit | |
multilinguality: | |
- monolingual | |
size_categories: | |
- 100K<n<1M | |
source_datasets: | |
- original | |
task_categories: | |
- text2text-generation | |
- text-generation | |
- fill-mask | |
task_ids: [] | |
paperswithcode_id: null | |
tags: | |
- code-modeling | |
- code-generation | |
dataset_info: | |
features: | |
- name: ast | |
sequence: | |
- name: type | |
dtype: string | |
- name: value | |
dtype: string | |
- name: children | |
sequence: int32 | |
config_name: ast | |
splits: | |
- name: train | |
num_bytes: 1870790180 | |
num_examples: 100000 | |
- name: test | |
num_bytes: 907514993 | |
num_examples: 50000 | |
download_size: 526642289 | |
dataset_size: 2778305173 | |
# Dataset Card for [py_ast] | |
## Table of Contents | |
- [Dataset Description](#dataset-description) | |
- [Dataset Summary](#dataset-summary) | |
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards) | |
- [Languages](#languages) | |
- [Dataset Structure](#dataset-structure) | |
- [Data Instances](#data-instances) | |
- [Data Fields](#data-fields) | |
- [Data Splits](#data-splits) | |
- [Dataset Creation](#dataset-creation) | |
- [Curation Rationale](#curation-rationale) | |
- [Source Data](#source-data) | |
- [Annotations](#annotations) | |
- [Personal and Sensitive Information](#personal-and-sensitive-information) | |
- [Considerations for Using the Data](#considerations-for-using-the-data) | |
- [Social Impact of Dataset](#social-impact-of-dataset) | |
- [Discussion of Biases](#discussion-of-biases) | |
- [Other Known Limitations](#other-known-limitations) | |
- [Additional Information](#additional-information) | |
- [Dataset Curators](#dataset-curators) | |
- [Licensing Information](#licensing-information) | |
- [Citation Information](#citation-information) | |
- [Contributions](#contributions) | |
## Dataset Description | |
- **homepage**: [py150](https://www.sri.inf.ethz.ch/py150) | |
- **Paper**: [Probabilistic Model for Code with Decision Trees](https://www.semanticscholar.org/paper/Probabilistic-model-for-code-with-decision-trees-Raychev-Bielik/62e176977d439aac2e2d7eca834a7a99016dfcaf) | |
- **Leaderboard:** | |
- **Point of Contact:** | |
### Dataset Summary | |
The dataset consists of parsed ASTs that were used to train and evaluate the DeepSyn tool. | |
The Python programs are collected from GitHub repositories | |
by removing duplicate files, removing project forks (copy of another existing repository), | |
keeping only programs that parse and have at most 30'000 nodes in the AST and | |
we aim to remove obfuscated files | |
### Supported Tasks and Leaderboards | |
Code Representation, Unsupervised Learning | |
### Languages | |
Python | |
## Dataset Structure | |
### Data Instances | |
A typical datapoint contains an AST of a python program, parsed. | |
The main key is `ast` wherein every program's AST is stored. | |
Each children would have, | |
`type` which will formulate the type of the node. | |
`children` which enumerates if a given node has children(non-empty list). | |
`value`, if the given node has any hardcoded value(else "N/A"). | |
An example would be, | |
''' | |
[ {"type":"Module","children":[1,4]},{"type":"Assign","children":[2,3]},{"type":"NameStore","value":"x"},{"type":"Num","value":"7"}, {"type":"Print","children":[5]}, {"type":"BinOpAdd","children":[6,7]}, {"type":"NameLoad","value":"x"}, {"type":"Num","value":"1"} ] | |
''' | |
### Data Fields | |
- `ast`: a list of dictionaries, wherein every dictionary is a node in the Abstract Syntax Tree. | |
- `type`: explains the type of the node. | |
- `children`: list of nodes which are children under the given | |
- `value`: hardcoded value, if the node holds an hardcoded value. | |
### Data Splits | |
The data is split into a training and test set. | |
The final split sizes are as follows: | |
| | train | validation | | |
|------------------|--------:|------------:| | |
| py_ast examples | 100000 | 50000 | | |
## Dataset Creation | |
[More Information Needed] | |
### Curation Rationale | |
[More Information Needed] | |
### Source Data | |
#### Initial Data Collection and Normalization | |
[More Information Needed] | |
#### Who are the source language producers? | |
[More Information Needed] | |
### Annotations | |
#### Annotation process | |
[More Information Needed] | |
#### Who are the annotators? | |
[More Information Needed] | |
### Personal and Sensitive Information | |
[More Information Needed] | |
## Considerations for Using the Data | |
### Social Impact of Dataset | |
[More Information Needed] | |
### Discussion of Biases | |
[More Information Needed] | |
### Other Known Limitations | |
[More Information Needed] | |
## Additional Information | |
### Dataset Curators | |
Raychev, V., Bielik, P., and Vechev, M | |
### Licensing Information | |
MIT, BSD and Apache | |
### Citation Information | |
@InProceedings{OOPSLA ’16, ACM, | |
title = {Probabilistic Model for Code with Decision Trees.}, | |
authors={Raychev, V., Bielik, P., and Vechev, M.}, | |
year={2016} | |
} | |
``` | |
@inproceedings{10.1145/2983990.2984041, | |
author = {Raychev, Veselin and Bielik, Pavol and Vechev, Martin}, | |
title = {Probabilistic Model for Code with Decision Trees}, | |
year = {2016}, | |
isbn = {9781450344449}, | |
publisher = {Association for Computing Machinery}, | |
address = {New York, NY, USA}, | |
url = {https://doi.org/10.1145/2983990.2984041}, | |
doi = {10.1145/2983990.2984041}, | |
booktitle = {Proceedings of the 2016 ACM SIGPLAN International Conference on Object-Oriented Programming, Systems, Languages, and Applications}, | |
pages = {731–747}, | |
numpages = {17}, | |
keywords = {Code Completion, Decision Trees, Probabilistic Models of Code}, | |
location = {Amsterdam, Netherlands}, | |
series = {OOPSLA 2016} | |
} | |
``` | |
### Contributions | |
Thanks to [@reshinthadithyan](https://github.com/reshinthadithyan) for adding this dataset. |