Thanks. iQ1_S Actually Works.

#28
by deleted - opened
deleted

Since your Mixtral-8x22B-v0.1.IQ1_S version just barely fit in my 32 GB of RAM I decided to give it a shot simply for the sake of curiosity and expecting nonsensical outputs, but it was surprisingly coherent and correctly answered some simple questions. How a 1-bit quantization functions at all is beyond me.

Thank you for confirming my own experience with IQ-1 models! This new way of quantizing by using imatix.data really improves the quality of the quantized models!

PS: this is a fine-tuned model by HuggingFace based on this mode, you might want to check the IQ-1 models I just pushed there as well. That should be much easier to instruct: https://huggingface.co/MaziyarPanahi/zephyr-orpo-141b-A35b-v0.1-GGUF

Sign up or log in to comment