Adding Evaluation Results

#2
Files changed (1) hide show
  1. README.md +14 -1
README.md CHANGED
@@ -164,4 +164,17 @@ The model uses a [sentencepiece](https://github.com/google/sentencepiece)-based
164
  ---
165
 
166
  # Licenese
167
- [The MIT license](https://opensource.org/licenses/MIT)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
164
  ---
165
 
166
  # Licenese
167
+ [The MIT license](https://opensource.org/licenses/MIT)
168
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
169
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_rinna__bilingual-gpt-neox-4b-8k)
170
+
171
+ | Metric | Value |
172
+ |-----------------------|---------------------------|
173
+ | Avg. | 27.67 |
174
+ | ARC (25-shot) | 28.58 |
175
+ | HellaSwag (10-shot) | 43.94 |
176
+ | MMLU (5-shot) | 25.38 |
177
+ | TruthfulQA (0-shot) | 47.48 |
178
+ | Winogrande (5-shot) | 47.99 |
179
+ | GSM8K (5-shot) | 0.0 |
180
+ | DROP (3-shot) | 0.31 |