Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Update app.py
Browse files
app.py
CHANGED
@@ -90,7 +90,7 @@ def search_arxiv(query):
|
|
90 |
st.title("▶️ Semantic and Episodic Memory System")
|
91 |
client = Client("awacke1/Arxiv-Paper-Search-And-QA-RAG-Pattern")
|
92 |
search_query = query
|
93 |
-
top_n_results = st.slider(key='topnresults', label="Top n results as context", min_value=4, max_value=100, value=100)
|
94 |
search_source = st.selectbox("Search Source", ["Semantic Search - up to 10 Mar 2024", "Arxiv Search - Latest - (EXPERIMENTAL)"])
|
95 |
llm_model = st.selectbox("LLM Model", ["mistralai/Mixtral-8x7B-Instruct-v0.1", "mistralai/Mistral-7B-Instruct-v0.2", "google/gemma-7b-it", "None"])
|
96 |
|
@@ -100,7 +100,7 @@ def search_arxiv(query):
|
|
100 |
|
101 |
result = client.predict(
|
102 |
search_query,
|
103 |
-
|
104 |
search_source,
|
105 |
llm_model,
|
106 |
api_name="/update_with_rag_md"
|
|
|
90 |
st.title("▶️ Semantic and Episodic Memory System")
|
91 |
client = Client("awacke1/Arxiv-Paper-Search-And-QA-RAG-Pattern")
|
92 |
search_query = query
|
93 |
+
#top_n_results = st.slider(key='topnresults', label="Top n results as context", min_value=4, max_value=100, value=100)
|
94 |
search_source = st.selectbox("Search Source", ["Semantic Search - up to 10 Mar 2024", "Arxiv Search - Latest - (EXPERIMENTAL)"])
|
95 |
llm_model = st.selectbox("LLM Model", ["mistralai/Mixtral-8x7B-Instruct-v0.1", "mistralai/Mistral-7B-Instruct-v0.2", "google/gemma-7b-it", "None"])
|
96 |
|
|
|
100 |
|
101 |
result = client.predict(
|
102 |
search_query,
|
103 |
+
100,
|
104 |
search_source,
|
105 |
llm_model,
|
106 |
api_name="/update_with_rag_md"
|