SiraH commited on
Commit
6305f5c
1 Parent(s): d66aee8

change batch size

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -159,8 +159,8 @@ class UploadDoc:
159
 
160
  return documents
161
 
162
- def split_docs(documents,chunk_size=500):
163
- text_splitter = RecursiveCharacterTextSplitter(chunk_size=chunk_size, chunk_overlap=100)
164
  sp_docs = text_splitter.split_documents(documents)
165
  return sp_docs
166
 
@@ -168,7 +168,7 @@ def split_docs(documents,chunk_size=500):
168
  def load_llama2_llamaCpp():
169
  core_model_name = "llama-2-7b-chat.Q4_0.gguf"
170
  #n_gpu_layers = 32
171
- n_batch = 512
172
  callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
173
  llm = LlamaCpp(
174
  model_path=core_model_name,
@@ -243,7 +243,7 @@ def main():
243
  text += page.extract_text()
244
  db = FAISS.from_texts(text, embeddings)
245
  qa_chain = ConversationalRetrievalChain(
246
- retriever =db.as_retriever(search_type="similarity_score_threshold", search_kwargs={'k':3, "score_threshold": 0.7}),
247
  question_generator=question_generator,
248
  #condense_question_prompt=CONDENSE_QUESTION_PROMPT,
249
  combine_docs_chain=doc_chain,
 
159
 
160
  return documents
161
 
162
+ def split_docs(documents,chunk_size=1000):
163
+ text_splitter = RecursiveCharacterTextSplitter(chunk_size=chunk_size, chunk_overlap=200)
164
  sp_docs = text_splitter.split_documents(documents)
165
  return sp_docs
166
 
 
168
  def load_llama2_llamaCpp():
169
  core_model_name = "llama-2-7b-chat.Q4_0.gguf"
170
  #n_gpu_layers = 32
171
+ n_batch = 8
172
  callback_manager = CallbackManager([StreamingStdOutCallbackHandler()])
173
  llm = LlamaCpp(
174
  model_path=core_model_name,
 
243
  text += page.extract_text()
244
  db = FAISS.from_texts(text, embeddings)
245
  qa_chain = ConversationalRetrievalChain(
246
+ retriever =db.as_retriever(search_type="similarity_score_threshold", search_kwargs={'k':2, "score_threshold": 0.7}),
247
  question_generator=question_generator,
248
  #condense_question_prompt=CONDENSE_QUESTION_PROMPT,
249
  combine_docs_chain=doc_chain,