Spaces:
Sleeping
Sleeping
change batch size
Browse files
app.py
CHANGED
@@ -159,8 +159,8 @@ class UploadDoc:
|
|
159 |
|
160 |
return documents
|
161 |
|
162 |
-
def split_docs(documents,chunk_size=
|
163 |
-
text_splitter = RecursiveCharacterTextSplitter(chunk_size=chunk_size, chunk_overlap=
|
164 |
sp_docs = text_splitter.split_documents(documents)
|
165 |
return sp_docs
|
166 |
|
@@ -168,7 +168,7 @@ def split_docs(documents,chunk_size=500):
|
|
168 |
def load_llama2_llamaCpp():
|
169 |
core_model_name = "llama-2-7b-chat.Q4_0.gguf"
|
170 |
#n_gpu_layers = 32
|
171 |
-
n_batch =
|
172 |
callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
|
173 |
llm = LlamaCpp(
|
174 |
model_path=core_model_name,
|
@@ -243,7 +243,7 @@ def main():
|
|
243 |
text += page.extract_text()
|
244 |
db = FAISS.from_texts(text, embeddings)
|
245 |
qa_chain = ConversationalRetrievalChain(
|
246 |
-
retriever =db.as_retriever(search_type="similarity_score_threshold", search_kwargs={'k':
|
247 |
question_generator=question_generator,
|
248 |
#condense_question_prompt=CONDENSE_QUESTION_PROMPT,
|
249 |
combine_docs_chain=doc_chain,
|
|
|
159 |
|
160 |
return documents
|
161 |
|
162 |
+
def split_docs(documents,chunk_size=1000):
|
163 |
+
text_splitter = RecursiveCharacterTextSplitter(chunk_size=chunk_size, chunk_overlap=200)
|
164 |
sp_docs = text_splitter.split_documents(documents)
|
165 |
return sp_docs
|
166 |
|
|
|
168 |
def load_llama2_llamaCpp():
|
169 |
core_model_name = "llama-2-7b-chat.Q4_0.gguf"
|
170 |
#n_gpu_layers = 32
|
171 |
+
n_batch = 8
|
172 |
callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
|
173 |
llm = LlamaCpp(
|
174 |
model_path=core_model_name,
|
|
|
243 |
text += page.extract_text()
|
244 |
db = FAISS.from_texts(text, embeddings)
|
245 |
qa_chain = ConversationalRetrievalChain(
|
246 |
+
retriever =db.as_retriever(search_type="similarity_score_threshold", search_kwargs={'k':2, "score_threshold": 0.7}),
|
247 |
question_generator=question_generator,
|
248 |
#condense_question_prompt=CONDENSE_QUESTION_PROMPT,
|
249 |
combine_docs_chain=doc_chain,
|