UltraLM-13b-v2.0 / README.md
leaderboard-pr-bot's picture
Adding Evaluation Results
321c491
metadata
license: mit

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 52.85
ARC (25-shot) 62.63
HellaSwag (10-shot) 81.49
MMLU (5-shot) 56.17
TruthfulQA (0-shot) 49.48
Winogrande (5-shot) 76.48
GSM8K (5-shot) 10.99
DROP (3-shot) 32.69