--- license: apache-2.0 library_name: peft tags: - trl - dpo - unsloth base_model: ayoubkirouane/Mistral-SLERP-Merged7B model-index: - name: outputs results: [] datasets: - HuggingFaceH4/ultrafeedback_binarized - ayoubkirouane/Orca-Direct-Preference-Optimization pipeline_tag: text-generation --- # Mistral-SLERP-Merged7B-DPO - DPO PEFT finetuned version from my [Mistral-SLERP-Merged7B](https://huggingface.co/ayoubkirouane/Mistral-SLERP-Merged7B) ## What is DPO ? ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6338c06c107c4835a05699f9/zIQWFZGxngGE8rQiodl2K.png) Direct Preference Optimization (DPO) is an algorithm introduced in order to achieve precise control of the behavior of large-scale unsupervised language models (LMs). It is a parameterization of the reward model in Reinforcement Learning from Human Feedback (RLHF) that enables the extraction of the corresponding optimal policy in closed form. This allows for the solution of the standard RLHF problem with only a simple classification loss. DPO eliminates the need for sampling from the LM during fine-tuning or performing significant hyperparameter tuning, making it stable, performant, and computationally lightweight. Experiments have shown that DPO can fine-tune LMs to align with human preferences as well as or better than existing methods. It has been found to be particularly effective in controlling the sentiment of generations and matching or improving response quality in summarization and single-turn dialogue.