awacke1 commited on
Commit
9054cad
1 Parent(s): 228f489

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -8
app.py CHANGED
@@ -152,7 +152,7 @@ def add_witty_humor_buttons():
152
  if col7[0].button("More Funny Rhymes 🎙️"):
153
  StreamLLMChatResponse(descriptions["More Funny Rhymes 🎙️"])
154
 
155
- def addDocumentHTML5(result):
156
  documentHTML5='''
157
  <!DOCTYPE html>
158
  <html>
@@ -180,7 +180,7 @@ def addDocumentHTML5(result):
180
  '''
181
 
182
  components.html(documentHTML5, width=1280, height=1024)
183
- return result
184
 
185
 
186
  # 3. Stream Llama Response
@@ -220,7 +220,7 @@ def StreamLLMChatResponse(prompt):
220
 
221
  except:
222
  st.write('Stream llm issue')
223
- #add_documentHTML5(result)
224
  return result
225
  except:
226
  st.write('Llama model is asleep. Starting up now on A10 - please give 5 minutes then retry as KEDA scales up from zero to activate running container(s).')
@@ -675,10 +675,8 @@ def main():
675
  if len(file_contents) > 0:
676
  if next_action=='open':
677
  file_content_area = st.text_area("File Contents:", file_contents, height=500)
678
- #addDocumentHTML5(file_contents)
679
  if next_action=='md':
680
  st.markdown(file_contents)
681
- #addDocumentHTML5(file_contents)
682
  if next_action=='search':
683
  file_content_area = st.text_area("File Contents:", file_contents, height=500)
684
  st.write('Reasoning with your inputs...')
@@ -687,9 +685,7 @@ def main():
687
  response = StreamLLMChatResponse(file_contents)
688
  filename = generate_filename(user_prompt, ".md")
689
  create_file(filename, file_contents, response, should_save)
690
-
691
- #addDocumentHTML5(file_contents)
692
- addDocumentHTML5(response)
693
 
694
  # old - gpt
695
  #response = chat_with_model(user_prompt, file_contents, model_choice)
 
152
  if col7[0].button("More Funny Rhymes 🎙️"):
153
  StreamLLMChatResponse(descriptions["More Funny Rhymes 🎙️"])
154
 
155
+ def SpeechSynthesis(result):
156
  documentHTML5='''
157
  <!DOCTYPE html>
158
  <html>
 
180
  '''
181
 
182
  components.html(documentHTML5, width=1280, height=1024)
183
+ #return result
184
 
185
 
186
  # 3. Stream Llama Response
 
220
 
221
  except:
222
  st.write('Stream llm issue')
223
+ SpeechSynthesis(result)
224
  return result
225
  except:
226
  st.write('Llama model is asleep. Starting up now on A10 - please give 5 minutes then retry as KEDA scales up from zero to activate running container(s).')
 
675
  if len(file_contents) > 0:
676
  if next_action=='open':
677
  file_content_area = st.text_area("File Contents:", file_contents, height=500)
 
678
  if next_action=='md':
679
  st.markdown(file_contents)
 
680
  if next_action=='search':
681
  file_content_area = st.text_area("File Contents:", file_contents, height=500)
682
  st.write('Reasoning with your inputs...')
 
685
  response = StreamLLMChatResponse(file_contents)
686
  filename = generate_filename(user_prompt, ".md")
687
  create_file(filename, file_contents, response, should_save)
688
+ SpeechSynthesis(response)
 
 
689
 
690
  # old - gpt
691
  #response = chat_with_model(user_prompt, file_contents, model_choice)