--- annotations_creators: - expert-generated language_creators: - expert-generated language: - code license: - mit multilinguality: - monolingual source_datasets: - original task_categories: - text2text-generation task_ids: [] pretty_name: OpenAI HumanEval-Infilling tags: - code-generation --- # HumanEval-Infilling ## Dataset Description - **Repository:** https://github.com/openai/human-eval-infilling - **Paper:** https://arxiv.org/pdf/2207.14255 ## Dataset Summary [HumanEval-Infilling](https://github.com/openai/human-eval-infilling) is a benchmark for infilling tasks, derived from [HumanEval](https://huggingface.co/datasets/openai_humaneval) benchmark for the evaluation of code generation models. ## Dataset Structure To load the dataset you need to specify a subset. By default `HumanEval-SingleLineInfilling` is loaded. ```python from datasets import load_dataset ds = load_dataset("humaneval_infilling", "HumanEval-RandomSpanInfilling") DatasetDict({ test: Dataset({ features: ['task_id', 'entry_point', 'prompt', 'suffix', 'canonical_solution', 'test'], num_rows: 1640 }) }) ``` ## Subsets This dataset has 4 subsets: HumanEval-MultiLineInfilling, HumanEval-SingleLineInfilling, HumanEval-RandomSpanInfilling, HumanEval-RandomSpanInfillingLight. The single-line, multi-line, random span infilling and its light version have 1033, 5815, 1640 and 164 tasks, respectively. ## Citation ``` @article{bavarian2022efficient, title={Efficient Training of Language Models to Fill in the Middle}, author={Bavarian, Mohammad and Jun, Heewoo and Tezak, Nikolas and Schulman, John and McLeavey, Christine and Tworek, Jerry and Chen, Mark}, journal={arXiv preprint arXiv:2207.14255}, year={2022} } ```