Adding Evaluation Results

#1
Files changed (1) hide show
  1. README.md +14 -1
README.md CHANGED
@@ -87,4 +87,17 @@ Please read this disclaimer carefully before using the large language model prov
87
 
88
  ---
89
  license: llama2
90
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
87
 
88
  ---
89
  license: llama2
90
+ ---
91
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
92
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_CobraMamba__mamba-gpt-7b)
93
+
94
+ | Metric | Value |
95
+ |-----------------------|---------------------------|
96
+ | Avg. | 43.78 |
97
+ | ARC (25-shot) | 51.19 |
98
+ | HellaSwag (10-shot) | 75.4 |
99
+ | MMLU (5-shot) | 47.47 |
100
+ | TruthfulQA (0-shot) | 42.06 |
101
+ | Winogrande (5-shot) | 71.67 |
102
+ | GSM8K (5-shot) | 11.98 |
103
+ | DROP (3-shot) | 6.72 |