Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
erikqu
/
Mixtral_7Bx4_MOE_24B-gguf
like
0
GGUF
Inference Endpoints
License:
mit
Model card
Files
Files and versions
Community
1
Deploy
Use this model
Original repo: https://huggingface.co/cloudyu/Mixtral_7Bx4_MOE_24B
This is the Q4_K_M version.
Model requires ~13G vram.
Original repo:
https://huggingface.co/cloudyu/Mixtral_7Bx4_MOE_24B
This is the Q4_K_M version.
Model requires ~13G vram.
Downloads last month
3
GGUF
Model size
24.2B params
Architecture
llama
4-bit
Q4_K_M
Inference API
Unable to determine this model's library. Check the
docs
.