Would like to add a quantitative model of GPTQ-IN8

#1
by warlock-edward - opened

Since my GPU is a 7th generation architecture and doesn't support AWQ model, expect a quantization model for int8 with GPTQ, thanks a lot!

Sign up or log in to comment