Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -8,7 +8,7 @@ tokenizer = AutoTokenizer.from_pretrained(checkpoint)
|
|
| 8 |
model = AutoModelForCausalLM.from_pretrained(checkpoint,torch_dtype="auto").to(device)
|
| 9 |
|
| 10 |
def predict(message, history):
|
| 11 |
-
history.append({"role": "user", "
|
| 12 |
input_text = tokenizer.apply_chat_template(history, tokenize=False)
|
| 13 |
inputs = tokenizer.encode(input_text, return_tensors="pt").to(device)
|
| 14 |
outputs = model.generate(inputs, max_new_tokens=3200, temperature=0.2, top_p=0.9, do_sample=True)
|
|
|
|
| 8 |
model = AutoModelForCausalLM.from_pretrained(checkpoint,torch_dtype="auto").to(device)
|
| 9 |
|
| 10 |
def predict(message, history):
|
| 11 |
+
history.append({"role": "user", "content": message})
|
| 12 |
input_text = tokenizer.apply_chat_template(history, tokenize=False)
|
| 13 |
inputs = tokenizer.encode(input_text, return_tensors="pt").to(device)
|
| 14 |
outputs = model.generate(inputs, max_new_tokens=3200, temperature=0.2, top_p=0.9, do_sample=True)
|