IQ3_XS quant not visible in sidebar

#1
by x0wllaar - opened

Hello! Thank you for the quants!
I've noticed that IQ3_XS is missing from the sidebar thing, making using the model with Ollama annoying and error-prone:

image.png

Is this a problem on my end? If not, if it fixable for yours?

Thank you!

it's not on yours or mine, it would be on the HF side, @reach-vb ?

Oh, ok!

I've honestly noticed IQ3_XS/IQ3_XXS missing from a lot of these sidebars in your repos, e.g. https://huggingface.co/bartowski/Qwen2.5-32B-Instruct-GGUF, https://huggingface.co/bartowski/Meta-Llama-3.1-70B-Instruct-GGUF, https://huggingface.co/bartowski/gemma-2-27b-it-GGUF (IQ3_XXS is visible, but IQ3_XS is not) (just 3 random ones)

It's weird. Thank you for looking into this

oh that's even stranger, i assumed they didn't show up anywhere !

Oh wait, I think IQ3_XS is missing everywhere, but IQ3_XXS shows up

I think IQ3_XS doesn't show up for other people either: https://huggingface.co/mradermacher/ThinkPhi1.1-Tensors-i1-GGUF, https://huggingface.co/mradermacher/Qwenvergence-14B-v11-i1-GGUF

Seems like HF has a strange aversion to IQ3_XS

Can you make an IQ3_XXS version of this so it shows up and 16GB VRAM people can use it? Thanks!

I personally do not understand the expediency of such quantization as IQ3_XXS. A model with fewer parameters, but in adequate quantization (>=Q4_K_M) will produce much better answers.

Sign up or log in to comment