Quantisation question?

#5
by pjw000 - opened

Using this model should I be able to create a GPTQ 4 bit version? My naive attempts using the official intruct model produce the following error which I have no idea how to deal with:

ValueError: Block pattern could not be match. Pass block_name_to_quantize argument in quantize_model

Specifically: I have no idea what the block name might be!

There are already people trying to get a GPTQ version. I'd suggest to have a look at this github discussion: https://github.com/AutoGPTQ/AutoGPTQ/issues/621

Thank you. That's a VERY helpful link! Much appreciated.

pjw000 changed discussion status to closed

Sign up or log in to comment