Spaces:
Build error
Build error
Update model.py
Browse files
model.py
CHANGED
@@ -13,7 +13,7 @@ os.system('pip install "unsloth[colab-new] @ git+https://github.com/unslothai/un
|
|
13 |
|
14 |
import gradio as gr
|
15 |
|
16 |
-
from transformers import TextStreamer
|
17 |
from unsloth import FastLanguageModel
|
18 |
from google.colab import drive
|
19 |
|
@@ -52,10 +52,10 @@ def chat_alpaca(message: str, history: list, temperature: float, max_new_tokens:
|
|
52 |
inputs = tokenizer([prompt], return_tensors="pt").to("cuda")
|
53 |
|
54 |
# Define the streamer
|
55 |
-
text_streamer = TextStreamer(tokenizer)
|
56 |
|
57 |
# Generate the response
|
58 |
-
outputs = model.generate(**inputs,
|
59 |
response = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]
|
60 |
|
61 |
# Return the response
|
|
|
13 |
|
14 |
import gradio as gr
|
15 |
|
16 |
+
# from transformers import TextStreamer
|
17 |
from unsloth import FastLanguageModel
|
18 |
from google.colab import drive
|
19 |
|
|
|
52 |
inputs = tokenizer([prompt], return_tensors="pt").to("cuda")
|
53 |
|
54 |
# Define the streamer
|
55 |
+
# text_streamer = TextStreamer(tokenizer)
|
56 |
|
57 |
# Generate the response
|
58 |
+
outputs = model.generate(**inputs, max_new_tokens=max_new_tokens, temperature=temperature)
|
59 |
response = tokenizer.batch_decode(outputs, skip_special_tokens=True)[0]
|
60 |
|
61 |
# Return the response
|