Edit model card

Model Card for Cerebrum-1.0-8x7b-GGUF

Quantized from https://huggingface.co/AetherResearch/Cerebrum-1.0-8x7b using llama.cpp commit 46acb3676718b983157058aecf729a2064fc7d34

Actual quants are currently uploading with my shitty german broadband speed of ~ 40Mbit/s, stay tuned.

Downloads last month
67
GGUF
Model size
46.7B params
Architecture
llama
Inference API
Input a message to start chatting with qwp4w3hyb/Cerebrum-1.0-8x7b-GGUF.
This model can be loaded on Inference API (serverless).