Edit model card

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 44.95
AI2 Reasoning Challenge (25-Shot) 37.80
HellaSwag (10-Shot) 59.73
MMLU (5-Shot) 45.80
TruthfulQA (0-shot) 42.28
Winogrande (5-shot) 60.22
GSM8k (5-shot) 23.88
Downloads last month
2,206
Safetensors
Model size
1.84B params
Tensor type
FP16
·

Dataset used to train vicgalle/OpenHermes-Qwen1.5-1.8B

Evaluation results