Edit model card

TinyLlama/TinyLlama-1.1B-Chat-v1.0 dpo finetuned on the argilla/distilabel-intel-orca-dpo-pairs dataset, which is the distilled version of https://huggingface.co/datasets/Intel/orca_dpo_pairs

GGUF Version: To be added Exllama Version: To be added

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 37.17
AI2 Reasoning Challenge (25-Shot) 36.18
HellaSwag (10-Shot) 61.15
MMLU (5-Shot) 25.09
TruthfulQA (0-shot) 38.05
Winogrande (5-shot) 60.85
GSM8k (5-shot) 1.67
Downloads last month
2,318
Safetensors
Model size
1.1B params
Tensor type
FP16
·

Dataset used to train eren23/DistiLabelOrca-TinyLLama-1.1B

Evaluation results