--- license: mit base_model: microsoft/phi-2 datasets: - TokenBender/code_instructions_122k_alpaca_style language: - en tags: - code - nlp --- ## Model Summary CodePhi2 is finetuning of the Microsoft Phi-2 LLM with **2.7 billion** parameters. It was finetuned on TokenBender's [code_instructions_122k_alpaca_style]("https://huggingface.co/datasets/TokenBender/code_instructions_122k_alpaca_style"). The end goal was to increase Phi-2's coding ability while imbuing the Alpaca format. ## Instruction Format (Alpaca) CodePhi2 has been finetuned on the Alpaca instruction format, and as such should be prompted like below: ``` Below is an instruction that describes a task. Write a response that appropriately completes the request. ### Instruction: {instruction} ### Response: ``` #### Notes If you are using transformers>=4.36.0, always load the model with trust_remote_code=True to prevent side-effects.