Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -69,7 +69,6 @@ def stream_chat(message: str, history: list, temperature: float, max_new_tokens:
|
|
69 |
|
70 |
input_ids = tokenizer.apply_chat_template(conversation, tokenize=True, add_generation_prompt=True, return_tensors="pt").to(model.device)
|
71 |
|
72 |
-
prompt_padded_len = len(input_ids[0])
|
73 |
|
74 |
gen_tokens= model.generate(
|
75 |
input_ids,
|
@@ -78,11 +77,7 @@ def stream_chat(message: str, history: list, temperature: float, max_new_tokens:
|
|
78 |
temperature=temperature,
|
79 |
)
|
80 |
|
81 |
-
|
82 |
-
gt[prompt_padded_len:] for gt in gen_tokens
|
83 |
-
]
|
84 |
-
|
85 |
-
gen_text = tokenizer.batch_decode(gen_tokens, skip_special_tokens=True)
|
86 |
|
87 |
return gen_text
|
88 |
|
|
|
69 |
|
70 |
input_ids = tokenizer.apply_chat_template(conversation, tokenize=True, add_generation_prompt=True, return_tensors="pt").to(model.device)
|
71 |
|
|
|
72 |
|
73 |
gen_tokens= model.generate(
|
74 |
input_ids,
|
|
|
77 |
temperature=temperature,
|
78 |
)
|
79 |
|
80 |
+
gen_text = tokenizer.decode(gen_tokens[0], skip_special_tokens=True)
|
|
|
|
|
|
|
|
|
81 |
|
82 |
return gen_text
|
83 |
|