--- license: apache-2.0 --- h2m/mhm-7b-v1.3-DPO-1 This is a DPO fine tuned mhm-7b-v1.3 on Intel/orca_dpo_pairs Model based on mistral. created using dare_ties and models from openllm leaderboard. Mixed 7 models into 1. 3 times merging. Just an experiment.