Text Generation
PEFT
Safetensors
trl
dpo
unsloth
conversational
ayoubkirouane commited on
Commit
9993da0
1 Parent(s): 5ec86cf

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -0
README.md CHANGED
@@ -18,6 +18,15 @@ pipeline_tag: text-generation
18
 
19
  - DPO finetuned version from my [Mistral-SLERP-Merged7B](https://huggingface.co/ayoubkirouane/Mistral-SLERP-Merged7B)
20
 
 
 
 
 
 
 
 
 
 
21
  ### Training hyperparameters
22
 
23
  The following hyperparameters were used during training:
 
18
 
19
  - DPO finetuned version from my [Mistral-SLERP-Merged7B](https://huggingface.co/ayoubkirouane/Mistral-SLERP-Merged7B)
20
 
21
+ ## What is DPO ?
22
+
23
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6338c06c107c4835a05699f9/zIQWFZGxngGE8rQiodl2K.png)
24
+
25
+ Direct Preference Optimization (DPO) is an algorithm introduced in order to achieve precise control of the behavior of large-scale unsupervised language models (LMs). It is a parameterization of the reward model in Reinforcement Learning from Human Feedback (RLHF) that enables the extraction of the corresponding optimal policy in closed form. This allows for the solution of the standard RLHF problem with only a simple classification loss.
26
+
27
+ DPO eliminates the need for sampling from the LM during fine-tuning or performing significant hyperparameter tuning, making it stable, performant, and computationally lightweight. Experiments have shown that DPO can fine-tune LMs to align with human preferences as well as or better than existing methods. It has been found to be particularly effective in controlling the sentiment of generations and matching or improving response quality in summarization and single-turn dialogue.
28
+
29
+
30
  ### Training hyperparameters
31
 
32
  The following hyperparameters were used during training: