Spaces:
Sleeping
Sleeping
Update chat_qwen.py
Browse files- chat_qwen.py +2 -2
chat_qwen.py
CHANGED
@@ -2,12 +2,12 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
2 |
|
3 |
def get_response(prompt: str):
|
4 |
model = AutoModelForCausalLM.from_pretrained(
|
5 |
-
"Qwen/
|
6 |
torch_dtype="auto",
|
7 |
device_map="auto",
|
8 |
)
|
9 |
|
10 |
-
tokenizer = AutoTokenizer.from_pretrained("Qwen/
|
11 |
|
12 |
prompt = "Give me a short introduction to large language model."
|
13 |
messages = [
|
|
|
2 |
|
3 |
def get_response(prompt: str):
|
4 |
model = AutoModelForCausalLM.from_pretrained(
|
5 |
+
"Qwen/Qwen2.5-32B-Instruct",
|
6 |
torch_dtype="auto",
|
7 |
device_map="auto",
|
8 |
)
|
9 |
|
10 |
+
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen2.5-32B-Instruct")
|
11 |
|
12 |
prompt = "Give me a short introduction to large language model."
|
13 |
messages = [
|