File size: 566 Bytes
253b348 38bbf30 253b348 38bbf30 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 |
---
license: apache-2.0
datasets:
- Intel/orca_dpo_pairs
---
# Model Summary
Neuralphi-2 is an experiment in DPO finetuning. It was made following Max Labonne's excellent [article](https://towardsdatascience.com/fine-tune-a-mistral-7b-model-with-direct-preference-optimization-708042745aac) about fine-tuning mistral-7b.
Neuralphi-2 is [phi-2-sft](https://huggingface.co/lxuechen/phi-2-sft) finetuned using DPO with [Intel/orca_dpo_pairs](https://huggingface.co/datasets/Intel/orca_dpo_pairs).
# Prompt Format
```
"""### Human: {instruction}
### Assistant:"""
```
|