Quantized models in GGUF

#4
by MaziyarPanahi - opened

Hi @v2ray

Thanks for converting and sharing this model, kudos! I am in the process of uploading GGUF models from 16bit all the way down to 2bit, for those with low resources:

https://huggingface.co/MaziyarPanahi/Mixtral-8x22B-v0.1-GGUF

Owner

owo

v2ray changed discussion status to closed
v2ray pinned discussion
v2ray changed discussion status to open

Hello all,

In LMStudio 0.2.18 : "llama.cpp error: 'illegal split file: 4, model must be loaded with the first split'" for the Q4
Any tip to solve this ?
Thank you.

Does LM Studio supports this new loading split?

Does LM Studio supports this new loading split?

It does support, I can confirm 😊

Sign up or log in to comment