Adding Evaluation Results

#19
Files changed (1) hide show
  1. README.md +14 -0
README.md CHANGED
@@ -107,3 +107,17 @@ Despite this, we have still worked hard to obtain opening the weights of the mod
107
  Our researchers have no authority to publicly release them without authorization.
108
 
109
  Thank you for your understanding.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
107
  Our researchers have no authority to publicly release them without authorization.
108
 
109
  Thank you for your understanding.
110
+
111
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
112
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_WizardLM__WizardLM-70B-V1.0)
113
+
114
+ | Metric | Value |
115
+ |-----------------------|---------------------------|
116
+ | Avg. | 57.17 |
117
+ | ARC (25-shot) | 65.44 |
118
+ | HellaSwag (10-shot) | 84.41 |
119
+ | MMLU (5-shot) | 64.05 |
120
+ | TruthfulQA (0-shot) | 54.81 |
121
+ | Winogrande (5-shot) | 80.82 |
122
+ | GSM8K (5-shot) | 17.97 |
123
+ | DROP (3-shot) | 32.71 |