Update README.md
Browse files
README.md
CHANGED
@@ -15,12 +15,6 @@ This repository provides a reproduction version of Tulu2-DPO-13B finetuned upon
|
|
15 |
|
16 |
## Performance
|
17 |
|
18 |
-
| Model | Size | Alignment | MT-Bench (score) | AlpacaEval (win rate %) |
|
19 |
-
|-------------|-----|----|---------------|--------------|
|
20 |
-
| **Tulu2-13b** | **13B** | **SFT** | **6.70** | **78.9** |
|
21 |
-
| **Tulu2-dpo-13b** | **13B** | **DPO** | **7.00** | **89.5** |
|
22 |
-
| **Reproduced-Tulu2-dpo-13b** | **13B** | **DPO** | **?** | **?** |
|
23 |
-
|
24 |
Check more progressive training metrics and final benchmark results in our [code repository](https://github.com/LuJunru/LLM_Finetune/tree/DPO).
|
25 |
|
26 |
## Input Format
|
|
|
15 |
|
16 |
## Performance
|
17 |
|
|
|
|
|
|
|
|
|
|
|
|
|
18 |
Check more progressive training metrics and final benchmark results in our [code repository](https://github.com/LuJunru/LLM_Finetune/tree/DPO).
|
19 |
|
20 |
## Input Format
|