CoNaLa / README.md
AhmedSSoliman's picture
Update README.md
c83ca9a
---
task_categories:
- Code Generation
- Translation
- Text2Text generation
---
# CoNaLa Dataset for Code Generation
## Table of content
- [Dataset Description](#dataset-description)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
## Dataset Descritpion
This dataset has been processed for Code Generation. CMU CoNaLa, the Code/Natural Language Challenge is a joint project of the Carnegie Mellon University NeuLab and STRUDEL Lab. This dataset was designed to test systems for generating program snippets from natural language. It is avilable at https://conala-corpus.github.io/ , and this is about 13k records from the full corpus of about 600k examples.
### Languages
English
## Dataset Structure
### Data Instances
A sample from this dataset looks as follows:
```json
[
{
"intent": "convert a list to a dictionary in python",
"snippet": "b = dict(zip(a[0::2], a[1::2]))"
},
{
"intent": "python - sort a list of nested lists",
"snippet": "l.sort(key=sum_nested)"
}
]
```
### Dataset Fields
The dataset has the following fields (also called "features"):
```json
{
"intent": "Value(dtype='string', id=None)",
"snippet": "Value(dtype='string', id=None)"
}
```
### Dataset Splits
This dataset is split into a train, validation and test split. The split sizes are as follow:
| Split name | Num samples |
| ------------ | ------------------- |
| train | 11125 |
| valid | 1237 |
| test | 500 |