ryanrwatkins commited on
Commit
9a1c32e
1 Parent(s): aa18bf6

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +33 -30
app.py CHANGED
@@ -104,32 +104,32 @@ def submit_message(prompt, prompt_template, temperature, max_tokens, context_len
104
  #vectordb = Chroma.from_documents(split_pages, embeddings, persist_directory=persist_directory)
105
  #vectordb.persist()
106
 
107
- path = './files'
108
- pdf_files = glob.glob(os.path.join(path, "*.pdf"))
109
 
110
- merger = PdfWriter()
111
 
112
  # add all file in the list to the merger object
113
- for pdf in pdf_files:
114
- merger.append(pdf)
115
- merger.write("merged-pdf.pdf")
116
- merger.close()
117
 
118
- reader = PdfReader("merged-pdf.pdf")
119
- raw_text = ''
120
- for i, page in enumerate(reader.pages):
121
- text = page.extract_text()
122
- if text:
123
- raw_text += text
124
- text_splitter = CharacterTextSplitter(
125
- separator = "\n",
126
- chunk_size = 1000,
127
- chunk_overlap = 200,
128
- length_function = len,
129
- )
130
- texts = text_splitter.split_text(raw_text)
131
- len(texts)
132
- embeddings = OpenAIEmbeddings()
133
 
134
 
135
  history = state['messages']
@@ -168,10 +168,13 @@ def submit_message(prompt, prompt_template, temperature, max_tokens, context_len
168
  #with open("foo.pkl", 'rb') as f:
169
  # new_docsearch = pickle.load(f)
170
 
171
- docsearch = FAISS.from_texts(texts, embeddings)
 
 
 
172
  #query = str(system_prompt + history[-context_length*2:] + [prompt_msg])
173
  query = str(system_prompt + history + [prompt_msg])
174
- docs = docsearch.similarity_search(query)
175
  #print(docs[0].page_content)
176
 
177
  chain = load_qa_chain(ChatOpenAI(temperature=temperature, max_tokens=max_tokens, model_name="gpt-3.5-turbo"), chain_type="stuff")
@@ -229,11 +232,11 @@ with gr.Blocks(css=css) as demo:
229
 
230
 
231
  with gr.Column(elem_id="col-container"):
232
- with open("embeddings.pkl", 'rb') as f:
233
- new_docsearch = pickle.load(f)
234
 
235
- query = str("performance")
236
- docs = new_docsearch.similarity_search(query)
237
 
238
  gr.Markdown("""# Chat with Needs Assessment Experts (Past and Present)
239
  ## Ask questions of experts on needs assessments, get responses from *needs assessment* version of ChatGPT.
@@ -246,8 +249,8 @@ with gr.Blocks(css=css) as demo:
246
  with gr.Row():
247
  with gr.Column():
248
  chatbot = gr.Chatbot(elem_id="chatbox")
249
- input_message = gr.Textbox(show_label=False, placeholder=docs, visible=True).style(container=False)
250
- #input_message = gr.Textbox(show_label=False, placeholder="Enter your needs assessment question and press enter", visible=True).style(container=False)
251
 
252
  btn_submit = gr.Button("Submit")
253
  total_tokens_str = gr.Markdown(elem_id="total_tokens_str")
 
104
  #vectordb = Chroma.from_documents(split_pages, embeddings, persist_directory=persist_directory)
105
  #vectordb.persist()
106
 
107
+ #path = './files'
108
+ #pdf_files = glob.glob(os.path.join(path, "*.pdf"))
109
 
110
+ #merger = PdfWriter()
111
 
112
  # add all file in the list to the merger object
113
+ #for pdf in pdf_files:
114
+ # merger.append(pdf)
115
+ #merger.write("merged-pdf.pdf")
116
+ #merger.close()
117
 
118
+ #reader = PdfReader("merged-pdf.pdf")
119
+ #raw_text = ''
120
+ #for i, page in enumerate(reader.pages):
121
+ # text = page.extract_text()
122
+ # if text:
123
+ # raw_text += text
124
+ #text_splitter = CharacterTextSplitter(
125
+ # separator = "\n",
126
+ # chunk_size = 1000,
127
+ # chunk_overlap = 200,
128
+ # length_function = len,
129
+ #)
130
+ #texts = text_splitter.split_text(raw_text)
131
+ #len(texts)
132
+ #embeddings = OpenAIEmbeddings()
133
 
134
 
135
  history = state['messages']
 
168
  #with open("foo.pkl", 'rb') as f:
169
  # new_docsearch = pickle.load(f)
170
 
171
+ #docsearch = FAISS.from_texts(texts, embeddings)
172
+ with open("embeddings.pkl", 'rb') as f:
173
+ new_docsearch = pickle.load(f)
174
+
175
  #query = str(system_prompt + history[-context_length*2:] + [prompt_msg])
176
  query = str(system_prompt + history + [prompt_msg])
177
+ docs = new_docsearch.similarity_search(query)
178
  #print(docs[0].page_content)
179
 
180
  chain = load_qa_chain(ChatOpenAI(temperature=temperature, max_tokens=max_tokens, model_name="gpt-3.5-turbo"), chain_type="stuff")
 
232
 
233
 
234
  with gr.Column(elem_id="col-container"):
235
+ #with open("embeddings.pkl", 'rb') as f:
236
+ # new_docsearch = pickle.load(f)
237
 
238
+ #query = str("performance")
239
+ #docs = new_docsearch.similarity_search(query)
240
 
241
  gr.Markdown("""# Chat with Needs Assessment Experts (Past and Present)
242
  ## Ask questions of experts on needs assessments, get responses from *needs assessment* version of ChatGPT.
 
249
  with gr.Row():
250
  with gr.Column():
251
  chatbot = gr.Chatbot(elem_id="chatbox")
252
+ #input_message = gr.Textbox(show_label=False, placeholder=docs, visible=True).style(container=False)
253
+ input_message = gr.Textbox(show_label=False, placeholder="Enter your needs assessment question and press enter", visible=True).style(container=False)
254
 
255
  btn_submit = gr.Button("Submit")
256
  total_tokens_str = gr.Markdown(elem_id="total_tokens_str")