vLLM Compatibility Issue with Unsloth's 4-bit Quantized Models - Shape Mismatch During Weight Loading

#1
by joserey1 - opened

just like in this discussion, using this model in vllm got this issue

image

tested by using nightly docker image
"vllm/vllm-openai@sha256:0d0104a260b69ce0bff9badde7620b8d796abfd067c327451f7ae0b09c761c9f"

Sign up or log in to comment