license: apache-2.0 | |
language: | |
- fr | |
- it | |
- de | |
- es | |
- en | |
tag: | |
- moe | |
- DPO | |
# Model Card for cloudyu/Mixtral-8x7B-Instruct-v0.1-DPO | |
* [DPO training try to improve mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1) | |
* [DPO Trainer](https://huggingface.co/docs/trl/main/en/dpo_trainer) | |
* metrics not test | |