Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
ayoubkirouane
/
Mistral-SLERP-Merged7B-DPO
like
0
Text Generation
PEFT
Safetensors
HuggingFaceH4/ultrafeedback_binarized
ayoubkirouane/Orca-Direct-Preference-Optimization
trl
dpo
unsloth
conversational
License:
apache-2.0
Model card
Files
Files and versions
Community
Use this model
main
Mistral-SLERP-Merged7B-DPO
/
tokenizer_config.json
Commit History
ayoubkirouane/Mistral-SLERP-Merged7B-DPO
4907099
verified
ayoubkirouane
commited on
Jan 24