Spaces:
Sleeping
Sleeping
RickMartel
commited on
Commit
•
6122c92
1
Parent(s):
b2bc435
Update app.py
Browse files
app.py
CHANGED
@@ -53,7 +53,7 @@ submit = form.form_submit_button('Submit')
|
|
53 |
|
54 |
if submit:
|
55 |
#generated = get_model_input(txt)
|
56 |
-
|
57 |
# output = model.generate(
|
58 |
# generated,
|
59 |
# stopping_criteria=stopping_criteria,
|
@@ -63,7 +63,7 @@ if submit:
|
|
63 |
# )
|
64 |
# out = tokenizer.decode(output[0], skip_special_tokens=True)
|
65 |
# out = out.replace('"', "")
|
66 |
-
|
67 |
input = tokenizer([tokenizer.bos_token + txt], return_tensors="pt")
|
68 |
streamer = TextIteratorStreamer( tokenizer )
|
69 |
generation_kwargs = dict(input, streamer=streamer,
|
@@ -77,4 +77,4 @@ if submit:
|
|
77 |
#generated_text += new_text
|
78 |
#print(new_text.replace('"', ""), end="")
|
79 |
#generated_text
|
80 |
-
st.write(new_text.replace('"', ""))
|
|
|
53 |
|
54 |
if submit:
|
55 |
#generated = get_model_input(txt)
|
56 |
+
with st.spinner('Processing...'):
|
57 |
# output = model.generate(
|
58 |
# generated,
|
59 |
# stopping_criteria=stopping_criteria,
|
|
|
63 |
# )
|
64 |
# out = tokenizer.decode(output[0], skip_special_tokens=True)
|
65 |
# out = out.replace('"', "")
|
66 |
+
#with st.expander("Response", expanded=True):
|
67 |
input = tokenizer([tokenizer.bos_token + txt], return_tensors="pt")
|
68 |
streamer = TextIteratorStreamer( tokenizer )
|
69 |
generation_kwargs = dict(input, streamer=streamer,
|
|
|
77 |
#generated_text += new_text
|
78 |
#print(new_text.replace('"', ""), end="")
|
79 |
#generated_text
|
80 |
+
st.write(new_text.replace('"', "").replace(tokenizer.bos_token,""))
|