No GGUF Quatization?

#1
by Goldenblood56 - opened

Is there a reason this does not get a GGUF quantization? Thanks for providing it in GPTQ I don't want to sound ungrateful. Thanks for the hard work TheBloke.

What does it take to GGUF export it

I didn't make GGUFs because I don't believe it's possible to use Llava with GGUF at this time. To get the image processing aspects, requires other components which are not supported in GGUF yet.

Actually llama.cpp/llava works with llava.gguf

Obsidian-3B-V0.5 - https://huggingface.co/nisten/obsidian-3b-multimodal-q6-gguf
Needs a llama.cpp fork tho (because StableLM-3B-4e1t base), instructions in model card.

Sign up or log in to comment