Improve installation + code snippets

#2
by Xenova HF staff - opened
Hugging Quants org
No description provided.
Hugging Quants org

Looks good to me! Should the dtype be bfloat16, or not when using AWQ?

Hugging Quants org

LGTM, thanks @Xenova ! And @pcuenq apparently for AWQ using float16 is recommended, if you set the torch_dtype=torch.bfloat16 the following message appears We suggest you to set torch_dtype=torch.float16 for better efficiency with AWQ., this is why we're using torch.float16 with AWQ :D

alvarobartt changed pull request status to merged

Sign up or log in to comment