Update README.md
Browse files
README.md
CHANGED
@@ -12,6 +12,9 @@ This repo contains a GGUF Quantized versions of the SFR-Iterative-DPO-Llama-3-8B
|
|
12 |
</a>
|
13 |
</div>
|
14 |
|
|
|
|
|
|
|
15 |
## Introduction
|
16 |
We release a state-of-the-art instruct model of its class, **SFR-Iterative-DPO-LLaMA-3-8B-R**.
|
17 |
On all three widely-used instruct model benchmarks: **Alpaca-Eval-V2**, **MT-Bench**, **Chat-Arena-Hard**, our model outperforms all models of similar size (e.g., LLaMA-3-8B-it), most large open-sourced models (e.g., Mixtral-8x7B-it),
|
|
|
12 |
</a>
|
13 |
</div>
|
14 |
|
15 |
+
weights from:
|
16 |
+
[maldv/SFR-Iterative-DPO-LLaMA-3-8B-R](https://huggingface.co/maldv/SFR-Iterative-DPO-LLaMA-3-8B-R)
|
17 |
+
|
18 |
## Introduction
|
19 |
We release a state-of-the-art instruct model of its class, **SFR-Iterative-DPO-LLaMA-3-8B-R**.
|
20 |
On all three widely-used instruct model benchmarks: **Alpaca-Eval-V2**, **MT-Bench**, **Chat-Arena-Hard**, our model outperforms all models of similar size (e.g., LLaMA-3-8B-it), most large open-sourced models (e.g., Mixtral-8x7B-it),
|