G-reen's picture
Update README.md
26b3272 verified
|
raw
history blame
No virus
948 Bytes

This model was trained as part of a series of experiments testing the performance of pure DPO vs SFT vs ORPO, all supported by Unsloth/Huggingface TRL.

Benchmarks

TBA

Training Details

Model: https://huggingface.co/unsloth/mistral-7b-v0.2-bnb-4bit

Dataset: https://huggingface.co/datasets/argilla/dpo-mix-7k

Rank: 8

Alpha: 16

Learning rate: 5e-6

Beta: 0.1

Batch size: 8

Epochs: 1

Learning rate scheduler: Linear

Prompt Format: You are a helpful assistant.<s>[INST] PROMPT [/INST]RESPONSE</s>

WanDB Reports image/png

image/png