Update README.md
Browse files
README.md
CHANGED
@@ -221,7 +221,7 @@ Also, we evaluate our model on the benchmarks of the first leaderboard using `li
|
|
221 |
| `FalconMamba-7B`<sup>*</sup> |62.03 | 80.82 | 62.11 | 73.64 | 53.42 | 52.54 | **64.09** |
|
222 |
| `TRI-ML/mamba-7b-rw`<sup>*</sup> | 51.25 | 80.85 | 33.41 | 71.11 | 23.13 | 4.70 | 44.03 |
|
223 |
|***Hybrid SSM-attention models***| | | | | | | |
|
224 |
-
| `recurrentgemma-9b
|
225 |
| `Zyphra/Zamba-7B-v1`<sup>*</sup> | 56.14 | 82.23 | 58.11 | 79.87 | 36.23 | 30.78 | 57.23 |
|
226 |
|***Transformer models*** | | | | | | | |
|
227 |
| `Falcon2-11B` | 59.73 | 82.91 | 58.37 | 78.30 | 52.56 | 53.83 | **64.28** |
|
@@ -229,7 +229,7 @@ Also, we evaluate our model on the benchmarks of the first leaderboard using `li
|
|
229 |
| `Mistral-7B-v0.1` | 59.98 | 83.31 | 64.16 | 78.37 | 42.15 | 37.83 | 60.97 |
|
230 |
| `gemma-7B` | 61.09 | 82.20 | 64.56 | 79.01 | 44.79 | 50.87 | 63.75 |
|
231 |
|
232 |
-
|
233 |
|
234 |
## Throughput
|
235 |
|
|
|
221 |
| `FalconMamba-7B`<sup>*</sup> |62.03 | 80.82 | 62.11 | 73.64 | 53.42 | 52.54 | **64.09** |
|
222 |
| `TRI-ML/mamba-7b-rw`<sup>*</sup> | 51.25 | 80.85 | 33.41 | 71.11 | 23.13 | 4.70 | 44.03 |
|
223 |
|***Hybrid SSM-attention models***| | | | | | | |
|
224 |
+
| `recurrentgemma-9b`<sup>**</sup> |52.00 | 80.40 | 60.50 | 73.60 | 38.60 | 42.60 | 57.95 |
|
225 |
| `Zyphra/Zamba-7B-v1`<sup>*</sup> | 56.14 | 82.23 | 58.11 | 79.87 | 36.23 | 30.78 | 57.23 |
|
226 |
|***Transformer models*** | | | | | | | |
|
227 |
| `Falcon2-11B` | 59.73 | 82.91 | 58.37 | 78.30 | 52.56 | 53.83 | **64.28** |
|
|
|
229 |
| `Mistral-7B-v0.1` | 59.98 | 83.31 | 64.16 | 78.37 | 42.15 | 37.83 | 60.97 |
|
230 |
| `gemma-7B` | 61.09 | 82.20 | 64.56 | 79.01 | 44.79 | 50.87 | 63.75 |
|
231 |
|
232 |
+
Mostly, we took evaluation results from both leaderboards. For the models marked by *star* we evaluated the tasks internally, while for the models marked by two *stars* the results were taken from paper or model card.
|
233 |
|
234 |
## Throughput
|
235 |
|