Edit model card

Model Card for Model ID

Optimum GPTQ quantized 8-bit version of Mixtral-8x22B-Instruct-v0.1
WIP - do not use currently
See original model card for more information.

How to load

Downloads last month
81
Safetensors
Model size
36.9B params
Tensor type
I32
·
FP16
·
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.