cassanof's picture
Update README.md
e41f85a verified
|
raw
history blame
1.64 kB
metadata
library_name: transformers
datasets:
  - nuprl/EditPackFT-Multi

EditCoder-Multi-6.7b (version 1) is a fine-tuned version of DeepSeek Coder (base model, 6.7b parameters) for instructional code editing. We utilize EditPackFT-Multi as our fine-tuning dataset. The model is trained on a variety of different languages.

More information can be found on our paper.

Citation

If you use our work, please cite our paper as such:

@misc{cassano2023edit,
      title={Can It Edit? Evaluating the Ability of Large Language Models to Follow Code Editing Instructions}, 
      author={Federico Cassano and Luisa Li and Akul Sethi and Noah Shinn and Abby Brennan-Jones and Anton Lozhkov and Carolyn Jane Anderson and Arjun Guha},
      year={2023},
      eprint={2312.12450},
      archivePrefix={arXiv},
      primaryClass={cs.SE}
}

Prompt

The model has been trained on the following prompt format:

## Code Before:
{before}
## Instruction:
{instruction}
## Code After:
{after}

Here is a python function that can be used for formatting the prompt correctly:

def edit_prompt(old, instr):
    before = f"""## Code Before:\n{old}\n"""
    instr = f"""## Instruction:\n{instr}\n"""
    after = f"""## Code After:\n"""
    return before + instr + after

Train Your Own EditCoder

We provide the full pipeline that was used for training our own edit-coder model. The pipeline and instructions can be found on our GitHub repository.