Edit model card

DPO Finetuned teknium/OpenHermes-2.5-Mistral-7B using argilla/distilabel-intel-orca-dpo-pairs.

Intel orca dpo pairs is a distilled version: https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs

of https://huggingface.co/datasets/Intel/orca_dpo_pairs

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 67.76
AI2 Reasoning Challenge (25-Shot) 65.87
HellaSwag (10-Shot) 84.78
MMLU (5-Shot) 63.65
TruthfulQA (0-shot) 54.24
Winogrande (5-shot) 78.22
GSM8k (5-shot) 59.82
Downloads last month
2,333
Safetensors
Model size
7.24B params
Tensor type
FP16
·

Dataset used to train eren23/DistilHermes-2.5-Mistral-7B

Evaluation results