mhm-7b-v1.3-DPO-1 / README.md
h2m's picture
Create README.md
6ebd98f verified
|
raw
history blame
No virus
258 Bytes
metadata
license: apache-2.0

h2m/mhm-7b-v1.3-DPO-1

This is a DPO fine tuned mhm-7b-v1.3 on Intel/orca_dpo_pairs

Model based on mistral. created using dare_ties and models from openllm leaderboard. Mixed 7 models into 1. 3 times merging.

Just an experiment.