Edit model card

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 76.10
AI2 Reasoning Challenge (25-Shot) 72.27
HellaSwag (10-Shot) 89.31
MMLU (5-Shot) 66.55
TruthfulQA (0-shot) 78.55
Winogrande (5-shot) 83.82
GSM8k (5-shot) 66.11
Downloads last month
4,483
Safetensors
Model size
10.7B params
Tensor type
FP16
·

Dataset used to train vicgalle/CarbonBeagle-11B-truthy

Evaluation results