Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

Mistral-7B-v0.1-DPO - bnb 8bits

Original model description:

license: apache-2.0

Mistral-7b-v0.1-DPO is a finetuned adapter from the original Mistral-7b model. In this adaptor, I am finetuning the LM head in addition to the regular modules that are normally finetuned. Below is the list of the finetuned modules: 'k_proj', 'gate_proj', 'v_proj', 'up_proj', 'q_proj', 'o_proj', 'down_proj', 'lm_head'

Downloads last month
8
Safetensors
Model size
7.24B params
Tensor type
F32
FP16
I8
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.