Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,7 @@
|
|
|
|
|
|
|
|
|
|
1 |
*This model was trained as part of a series of experiments testing the performance of pure DPO vs SFT vs ORPO, all supported by Unsloth/Huggingface TRL.*
|
2 |
|
3 |
**Benchmarks**
|
@@ -32,8 +36,4 @@ Prompt Format: ```You are a helpful assistant.<s>[INST] PROMPT [/INST]RESPONSE</
|
|
32 |
|
33 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65a5c0e82823ba72ed2cee7d/8DQ0WiypkVIJeK_Y18Wv0.png)
|
34 |
|
35 |
-
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
36 |
-
|
37 |
-
---
|
38 |
-
license: "apache 2.0"
|
39 |
-
---
|
|
|
1 |
+
---
|
2 |
+
license: "apache-2.0"
|
3 |
+
---
|
4 |
+
|
5 |
*This model was trained as part of a series of experiments testing the performance of pure DPO vs SFT vs ORPO, all supported by Unsloth/Huggingface TRL.*
|
6 |
|
7 |
**Benchmarks**
|
|
|
36 |
|
37 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65a5c0e82823ba72ed2cee7d/8DQ0WiypkVIJeK_Y18Wv0.png)
|
38 |
|
39 |
+
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
|
|
|
|
|
|
|