Update app.py
Browse files
app.py
CHANGED
@@ -31,19 +31,6 @@ tokens = model.generate(
|
|
31 |
do_sample=True,
|
32 |
stopping_criteria=StoppingCriteriaList([StopOnTokens()])
|
33 |
)
|
34 |
-
|
35 |
-
|
36 |
-
gr.Markdown("## StableLM-Tuned-Alpha-7b Chat")
|
37 |
-
gr.HTML('''<center><a href="https://huggingface.co/spaces/stabilityai/stablelm-tuned-alpha-chat?duplicate=true"><img src="https://bit.ly/3gLdBN6" alt="Duplicate Space"></a>Duplicate the Space to skip the queue and run in a private space</center>''')
|
38 |
-
chatbot = gr.Chatbot().style(height=500)
|
39 |
-
with gr.Row():
|
40 |
-
with gr.Column(scale=0.70):
|
41 |
-
msg = gr.Textbox(label="Chat Message Box", placeholder="Chat Message Box", show_label=False).style(container=False)
|
42 |
-
with gr.Column(scale=0.30):
|
43 |
-
with gr.Row():
|
44 |
-
submit = gr.Button("Submit")
|
45 |
-
clear = gr.Button("Clear")
|
46 |
-
system_msg = gr.Textbox(
|
47 |
-
start_message, label="System Message", interactive=False, visible=False)
|
48 |
|
49 |
print(tokenizer.decode(tokens[0], skip_special_tokens=True))
|
|
|
31 |
do_sample=True,
|
32 |
stopping_criteria=StoppingCriteriaList([StopOnTokens()])
|
33 |
)
|
34 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
35 |
|
36 |
print(tokenizer.decode(tokens[0], skip_special_tokens=True))
|