Edit model card

Model Summary

Neuralphi-2 is an experiment in DPO finetuning. It was made following Max Labonne's excellent article about fine-tuning mistral-7b. Neuralphi-2 is phi-2-sft finetuned using DPO with Intel/orca_dpo_pairs.

Prompt Format

"""### Human: {instruction}

### Assistant:"""
Downloads last month
1
Safetensors
Model size
2.78B params
Tensor type
FP16
·

Dataset used to train xz56/neuralphi-2