cloudyu's picture
Update README.md
cdfa5b3 verified
|
raw
history blame
362 Bytes
---
license: apache-2.0
language:
- fr
- it
- de
- es
- en
tag:
- moe
- DPO
---
# Model Card for cloudyu/Mixtral-8x7B-Instruct-v0.1-DPO
* [DPO training try to improve mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1)
* [DPO Trainer](https://huggingface.co/docs/trl/main/en/dpo_trainer)
* metrics not test