Update app.py
Browse files
app.py
CHANGED
|
@@ -1,14 +1,3 @@
|
|
| 1 |
-
# import subprocess
|
| 2 |
-
|
| 3 |
-
# sub_p_res = subprocess.run(['pip', 'install', 'langchain', 'sentence-transformers', 'transformers', 'faiss-gpu', 'PyPDF2',"gradio_client"], stdout=subprocess.PIPE).stdout.decode('utf-8') #<cc-cm>
|
| 4 |
-
# print("pip install downloded ", sub_p_res)
|
| 5 |
-
|
| 6 |
-
|
| 7 |
-
# command = 'CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 pip install llama-cpp-python'
|
| 8 |
-
|
| 9 |
-
# sub_p_res = subprocess.run(command, shell=True, check=True)
|
| 10 |
-
|
| 11 |
-
# print("llama-cpp-python GPU downloaded ",sub_p_res)
|
| 12 |
|
| 13 |
from gradio_client import Client
|
| 14 |
from langchain.document_loaders.text import TextLoader
|
|
@@ -25,7 +14,6 @@ from langchain.chains import ConversationalRetrievalChain
|
|
| 25 |
from huggingface_hub import hf_hub_download
|
| 26 |
from langchain.llms import LlamaCpp
|
| 27 |
from langchain.chains import LLMChain
|
| 28 |
-
|
| 29 |
import time
|
| 30 |
import streamlit as st
|
| 31 |
|
|
@@ -191,10 +179,10 @@ def generate_llm_response(prompt_input):
|
|
| 191 |
# User-provided prompt
|
| 192 |
if prompt := st.chat_input("What is up?"):
|
| 193 |
# Translate the user's input from Darija to English
|
| 194 |
-
if selected_language == "Darija":
|
| 195 |
-
|
| 196 |
-
else:
|
| 197 |
-
|
| 198 |
|
| 199 |
st.session_state.messages.append({"role": "user", "content": translated_prompt})
|
| 200 |
with st.chat_message("user", avatar="user.png"):
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
|
| 2 |
from gradio_client import Client
|
| 3 |
from langchain.document_loaders.text import TextLoader
|
|
|
|
| 14 |
from huggingface_hub import hf_hub_download
|
| 15 |
from langchain.llms import LlamaCpp
|
| 16 |
from langchain.chains import LLMChain
|
|
|
|
| 17 |
import time
|
| 18 |
import streamlit as st
|
| 19 |
|
|
|
|
| 179 |
# User-provided prompt
|
| 180 |
if prompt := st.chat_input("What is up?"):
|
| 181 |
# Translate the user's input from Darija to English
|
| 182 |
+
#if selected_language == "Darija":
|
| 183 |
+
# translated_prompt = translate(prompt, "Moroccan Darija", "English")
|
| 184 |
+
# else:
|
| 185 |
+
translated_prompt = prompt
|
| 186 |
|
| 187 |
st.session_state.messages.append({"role": "user", "content": translated_prompt})
|
| 188 |
with st.chat_message("user", avatar="user.png"):
|