natolambert
commited on
Commit
β’
c7eaf1c
1
Parent(s):
bc6c48f
Update README.md
Browse files
README.md
CHANGED
@@ -43,12 +43,12 @@ At the time of release, the Tulu-v2-dpo-70b model is approximately equal to GPT4
|
|
43 |
All smaller DPO'd models have strong performance per model size in the category and with lower verbosity (average completion length).
|
44 |
| Model | Size | Alignment | MT-Bench (score) | AlpacaEval (win rate %) |
|
45 |
|-------------|-----|----|---------------|--------------|
|
46 |
-
| **Tulu-v2-7b**
|
47 |
-
| **Tulu-v2-13b**
|
48 |
-
| **Tulu-v2-70b**
|
49 |
-
| **Tulu-v2-dpo-7b**
|
50 |
-
| **Tulu-v2-dpo-13b**
|
51 |
-
| **Tulu-v2-dpo-70b**
|
52 |
| StableLM-Tuned-Ξ± | 7B| dSFT |2.75| -|
|
53 |
| MPT-Chat | 7B |dSFT |5.42| -|
|
54 |
| Xwin-LMv0.1 | 7B| dPPO| 6.19| 87.83|
|
|
|
43 |
All smaller DPO'd models have strong performance per model size in the category and with lower verbosity (average completion length).
|
44 |
| Model | Size | Alignment | MT-Bench (score) | AlpacaEval (win rate %) |
|
45 |
|-------------|-----|----|---------------|--------------|
|
46 |
+
| **Tulu-v2-7b** πͺ | **7B** | **dDPO** | **TODO** | **TODO** |
|
47 |
+
| **Tulu-v2-13b** πͺ | **13B** | **dDPO** | **TODO** | **TODO** |
|
48 |
+
| **Tulu-v2-70b** πͺ | **70B** | **dDPO** | **TODO** | **TODO** |
|
49 |
+
| **Tulu-v2-dpo-7b** πͺ | **7B** | **dDPO** | **TODO** | **TODO** |
|
50 |
+
| **Tulu-v2-dpo-13b** πͺ | **13B** | **dDPO** | **TODO** | **TODO** |
|
51 |
+
| **Tulu-v2-dpo-70b** πͺ | **70B** | **dDPO** | **TODO** | **TODO** |
|
52 |
| StableLM-Tuned-Ξ± | 7B| dSFT |2.75| -|
|
53 |
| MPT-Chat | 7B |dSFT |5.42| -|
|
54 |
| Xwin-LMv0.1 | 7B| dPPO| 6.19| 87.83|
|