Adding Evaluation Results

#6
Files changed (1) hide show
  1. README.md +14 -1
README.md CHANGED
@@ -112,4 +112,17 @@ To cite this model:
112
  month = April
113
  }
114
  ```
115
- [![Model architecture](https://img.shields.io/badge/Model%20Arch-Transformer%20Decoder-green)](#model-architecture)|[![Model size](https://img.shields.io/badge/Params-5B-green)](#model-architecture)|[![Language](https://img.shields.io/badge/Language-en--US-lightgrey#model-badge)](#datasets)|![AUR license](https://img.shields.io/badge/license-Apache%202-blue)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
112
  month = April
113
  }
114
  ```
115
+ [![Model architecture](https://img.shields.io/badge/Model%20Arch-Transformer%20Decoder-green)](#model-architecture)|[![Model size](https://img.shields.io/badge/Params-5B-green)](#model-architecture)|[![Language](https://img.shields.io/badge/Language-en--US-lightgrey#model-badge)](#datasets)|![AUR license](https://img.shields.io/badge/license-Apache%202-blue)
116
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
117
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Writer__camel-5b-hf)
118
+
119
+ | Metric | Value |
120
+ |-----------------------|---------------------------|
121
+ | Avg. | 33.7 |
122
+ | ARC (25-shot) | 35.15 |
123
+ | HellaSwag (10-shot) | 57.62 |
124
+ | MMLU (5-shot) | 26.07 |
125
+ | TruthfulQA (0-shot) | 40.65 |
126
+ | Winogrande (5-shot) | 61.01 |
127
+ | GSM8K (5-shot) | 0.38 |
128
+ | DROP (3-shot) | 15.0 |