Edit model card

This model was trained as part of a series of experiments testing the performance of pure DPO vs SFT vs ORPO, all supported by Unsloth/Huggingface TRL.

Benchmarks

Average 59.62

ARC 59.39

HellaSwag 82.48

MMLU 62.61

TruthfulQA 40.38

Winogrande 78.37

GSM8K 34.5

Training Details

Duration: ~9 hours on one Kaggle T4 with Unsloth

Model: https://huggingface.co/unsloth/mistral-7b-v0.2-bnb-4bit

Dataset: https://huggingface.co/datasets/argilla/dpo-mix-7k

Rank: 8

Alpha: 16

Learning rate: 5e-5

Beta: 0.1

Batch size: 8

Epochs: 1

Learning rate scheduler: Linear

Prompt Format: ChatML

<|im_start|>system
You are a helpful assistant.<|im_end|>
<|im_start|>user
Why is the sky blue?<|im_end|>
<|im_start|>assistant

WanDB Reports

image/png

image/png

image/png

Downloads last month
1,273
Safetensors
Model size
3.86B params
Tensor type
F32
FP16
U8