Model Description

This is a 4bit GPTQ quantization of Llasa-3B by the HKUSTAudio team. I tested using a script written by GitHub user nivibilla, linked below. For some reason, I was not able to run it on my RTX 3090, while quantized Llasa-1B worked fine. Please let me know if you can get it working.

Model Sources

Downloads last month
5
Safetensors
Model size
973M params
Tensor type
I32
·
FP16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for AgeOfAlgorithms/Llasa-3b-GPTQ-4bit

Quantized
(6)
this model