Update README.md
Browse files
README.md
CHANGED
@@ -18,6 +18,19 @@ This is my first English & Chinese MoE Model based on
|
|
18 |
* [SUSTech/SUS-Chat-34B]
|
19 |
|
20 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
gpu code example
|
22 |
|
23 |
```
|
|
|
18 |
* [SUSTech/SUS-Chat-34B]
|
19 |
|
20 |
|
21 |
+
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
22 |
+
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_cloudyu__Mixtral_34Bx2_MoE_60B)
|
23 |
+
|
24 |
+
| Metric |Value|
|
25 |
+
|---------------------------------|----:|
|
26 |
+
|Avg. |76.66|
|
27 |
+
|AI2 Reasoning Challenge (25-Shot)|71.33|
|
28 |
+
|HellaSwag (10-Shot) |85.25|
|
29 |
+
|MMLU (5-Shot) |77.34|
|
30 |
+
|TruthfulQA (0-shot) |66.59|
|
31 |
+
|Winogrande (5-shot) |84.85|
|
32 |
+
|GSM8k (5-shot) |74.60|
|
33 |
+
|
34 |
gpu code example
|
35 |
|
36 |
```
|