Unable to run the model properly

#8
by zaz423129 - opened

code:
from transformers import AutoModelForCausalLM, AutoTokenizer
device = "cuda" # the device to load the model onto

model = AutoModelForCausalLM.from_pretrained(
"/home/llmuser/LLM_Z/my_projects/Qwen2-7B-Instruct",
torch_dtype="auto",
device_map="auto"
)
tokenizer = AutoTokenizer.from_pretrained("/home/llmuser/LLM_Z/my_projects/Qwen2-7B-Instruct")

prompt = "Give me a short introduction to large language model."
messages = [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": prompt}
]
text = tokenizer.apply_chat_template(
messages,
tokenize=False,
add_generation_prompt=True
)
model_inputs = tokenizer([text], return_tensors="pt").to(device)

generated_ids = model.generate(
model_inputs.input_ids,
max_new_tokens=512
)
generated_ids = [
output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
]

response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]

The error is as follows
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
Traceback (most recent call last):
File "/home/llmuser/LLaMA-Factory/Qwen-7B.py", line 23, in
generated_ids = model.generate(
File "/home/llmuser/anaconda3/envs/fine-tuning/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "/home/llmuser/anaconda3/envs/fine-tuning/lib/python3.10/site-packages/transformers/generation/utils.py", line 1758, in generate
result = self._sample(
File "/home/llmuser/anaconda3/envs/fine-tuning/lib/python3.10/site-packages/transformers/generation/utils.py", line 2410, in _sample
next_token_scores = logits_processor(input_ids, next_token_logits)
File "/home/llmuser/anaconda3/envs/fine-tuning/lib/python3.10/site-packages/transformers/generation/logits_process.py", line 98, in call
scores = processor(input_ids, scores)
File "/home/llmuser/anaconda3/envs/fine-tuning/lib/python3.10/site-packages/transformers/generation/logits_process.py", line 341, in call
score = torch.gather(scores, 1, input_ids)
RuntimeError: CUDA error: device-side assert triggered
CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
Compile with TORCH_USE_CUDA_DSA to enable device-side assertions.

image.png

Sign up or log in to comment