Some weights of the model checkpoint at Llama-2-7B-Chat-GPTQ were not used when initializing LlamaForCausalLM

#35
by thlw - opened

hello,i wonder why i use Automodel.from_pretrained and have some logs like Some weights of the model checkpoint at Llama-2-7B-Chat-GPTQ were not used when initializing LlamaForCausalLM. But the GPU use is about 8G,my huggingface is 4.40.2

Sign up or log in to comment