robinsmits
commited on
Commit
•
0705fd3
1
Parent(s):
7fb67dc
Update README.md
Browse files
README.md
CHANGED
@@ -17,7 +17,7 @@ inference: false
|
|
17 |
|
18 |
## Model description
|
19 |
|
20 |
-
This DPO aligned model is the merged version of the adapter model [robinsmits/Qwen1.5-7B-Dutch-Chat-Dpo](robinsmits/Qwen1.5-7B-Dutch-Chat-Dpo).
|
21 |
|
22 |
DPO Finetuning was performed on the Dutch [BramVanroy/ultra_feedback_dutch_cleaned](https://huggingface.co/datasets/BramVanroy/ultra_feedback_dutch_cleaned) dataset.
|
23 |
|
|
|
17 |
|
18 |
## Model description
|
19 |
|
20 |
+
This DPO aligned model is the merged version of the adapter model [robinsmits/Qwen1.5-7B-Dutch-Chat-Dpo](https://huggingface.co/robinsmits/Qwen1.5-7B-Dutch-Chat-Dpo).
|
21 |
|
22 |
DPO Finetuning was performed on the Dutch [BramVanroy/ultra_feedback_dutch_cleaned](https://huggingface.co/datasets/BramVanroy/ultra_feedback_dutch_cleaned) dataset.
|
23 |
|