Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -6,11 +6,11 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStream
|
|
6 |
import gradio as gr
|
7 |
from threading import Thread
|
8 |
|
9 |
-
MODEL_LIST = ["
|
10 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
11 |
-
MODEL = "
|
12 |
|
13 |
-
TITLE = "<h1><center>
|
14 |
|
15 |
PLACEHOLDER = """
|
16 |
<center>
|
@@ -83,7 +83,7 @@ def stream_chat(
|
|
83 |
top_k = top_k,
|
84 |
temperature = temperature,
|
85 |
repetition_penalty=penalty,
|
86 |
-
eos_token_id=
|
87 |
streamer=streamer,
|
88 |
)
|
89 |
|
@@ -110,9 +110,7 @@ with gr.Blocks(css=CSS, theme="soft") as demo:
|
|
110 |
additional_inputs=[
|
111 |
gr.Textbox(
|
112 |
value="""
|
113 |
-
You are
|
114 |
-
When faced with a user's request, you first engage in a lengthy and in-depth thinking process to explore possible solutions to the problem.
|
115 |
-
After completing your thoughts, you then provide a detailed explanation of the solution process in your response.
|
116 |
""",
|
117 |
label="System Prompt",
|
118 |
lines=5,
|
|
|
6 |
import gradio as gr
|
7 |
from threading import Thread
|
8 |
|
9 |
+
MODEL_LIST = ["CohereForAI/aya-expanse-8b"]
|
10 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
11 |
+
MODEL = "CohereForAI/aya-expanse-8b"
|
12 |
|
13 |
+
TITLE = "<h1><center>aya-expanse-8b</center></h1>"
|
14 |
|
15 |
PLACEHOLDER = """
|
16 |
<center>
|
|
|
83 |
top_k = top_k,
|
84 |
temperature = temperature,
|
85 |
repetition_penalty=penalty,
|
86 |
+
eos_token_id=255001,
|
87 |
streamer=streamer,
|
88 |
)
|
89 |
|
|
|
110 |
additional_inputs=[
|
111 |
gr.Textbox(
|
112 |
value="""
|
113 |
+
You are a helpful assistant.
|
|
|
|
|
114 |
""",
|
115 |
label="System Prompt",
|
116 |
lines=5,
|