Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
ayoubkirouane
/
Mistral-SLERP-Merged7B-DPO
like
0
Text Generation
PEFT
Safetensors
HuggingFaceH4/ultrafeedback_binarized
ayoubkirouane/Orca-Direct-Preference-Optimization
trl
dpo
unsloth
conversational
License:
apache-2.0
Model card
Files
Files and versions
Community
Deploy
Use this model
main
Mistral-SLERP-Merged7B-DPO
Commit History
Update README.md
8668468
verified
ayoubkirouane
commited on
Jan 24
Update README.md
9f0636c
verified
ayoubkirouane
commited on
Jan 24
Update README.md
9993da0
verified
ayoubkirouane
commited on
Jan 24
Update README.md
5ec86cf
verified
ayoubkirouane
commited on
Jan 24
Delete runs
2388e61
verified
ayoubkirouane
commited on
Jan 24
ayoubkirouane/Mistral-SLERP-Merged7B-DPO
4907099
verified
ayoubkirouane
commited on
Jan 24
initial commit
61ead37
verified
ayoubkirouane
commited on
Jan 24