Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
comaniac
/
Mixtral-8x22B-Instruct-v0.1-FP8-v2
like
0
Text Generation
Transformers
Safetensors
mixtral
conversational
Inference Endpoints
text-generation-inference
fp8
Model card
Files
Files and versions
Community
Train
Deploy
Use this model
Edit model card
YAML Metadata Warning:
empty or missing yaml metadata in repo card (
https://huggingface.co/docs/hub/model-cards#model-card-metadata
)
Mixtral-8x22B-Instruct-v0.1-FP8-v2
Evaluation
Mixtral-8x22B-Instruct-v0.1-FP8-v2
Weights and activations are per-tensor quantized to float8_e4m3.
Quantization with AutoFP8 with updated activation scaling factor names.
Calibration dataset: Ultrachat (mgoin/ultrachat_2k)
Samples: 2048
Sequence length: 8192
Evaluation
TBA
Downloads last month
9
Safetensors
Model size
141B params
Tensor type
BF16
·
F8_E4M3
·