Spaces:
Running
Running
ryanrwatkins
commited on
Commit
•
9a1c32e
1
Parent(s):
aa18bf6
Update app.py
Browse files
app.py
CHANGED
@@ -104,32 +104,32 @@ def submit_message(prompt, prompt_template, temperature, max_tokens, context_len
|
|
104 |
#vectordb = Chroma.from_documents(split_pages, embeddings, persist_directory=persist_directory)
|
105 |
#vectordb.persist()
|
106 |
|
107 |
-
path = './files'
|
108 |
-
pdf_files = glob.glob(os.path.join(path, "*.pdf"))
|
109 |
|
110 |
-
merger = PdfWriter()
|
111 |
|
112 |
# add all file in the list to the merger object
|
113 |
-
for pdf in pdf_files:
|
114 |
-
|
115 |
-
merger.write("merged-pdf.pdf")
|
116 |
-
merger.close()
|
117 |
|
118 |
-
reader = PdfReader("merged-pdf.pdf")
|
119 |
-
raw_text = ''
|
120 |
-
for i, page in enumerate(reader.pages):
|
121 |
-
|
122 |
-
|
123 |
-
|
124 |
-
text_splitter = CharacterTextSplitter(
|
125 |
-
|
126 |
-
|
127 |
-
|
128 |
-
|
129 |
-
)
|
130 |
-
texts = text_splitter.split_text(raw_text)
|
131 |
-
len(texts)
|
132 |
-
embeddings = OpenAIEmbeddings()
|
133 |
|
134 |
|
135 |
history = state['messages']
|
@@ -168,10 +168,13 @@ def submit_message(prompt, prompt_template, temperature, max_tokens, context_len
|
|
168 |
#with open("foo.pkl", 'rb') as f:
|
169 |
# new_docsearch = pickle.load(f)
|
170 |
|
171 |
-
docsearch = FAISS.from_texts(texts, embeddings)
|
|
|
|
|
|
|
172 |
#query = str(system_prompt + history[-context_length*2:] + [prompt_msg])
|
173 |
query = str(system_prompt + history + [prompt_msg])
|
174 |
-
docs =
|
175 |
#print(docs[0].page_content)
|
176 |
|
177 |
chain = load_qa_chain(ChatOpenAI(temperature=temperature, max_tokens=max_tokens, model_name="gpt-3.5-turbo"), chain_type="stuff")
|
@@ -229,11 +232,11 @@ with gr.Blocks(css=css) as demo:
|
|
229 |
|
230 |
|
231 |
with gr.Column(elem_id="col-container"):
|
232 |
-
with open("embeddings.pkl", 'rb') as f:
|
233 |
-
|
234 |
|
235 |
-
query = str("performance")
|
236 |
-
docs = new_docsearch.similarity_search(query)
|
237 |
|
238 |
gr.Markdown("""# Chat with Needs Assessment Experts (Past and Present)
|
239 |
## Ask questions of experts on needs assessments, get responses from *needs assessment* version of ChatGPT.
|
@@ -246,8 +249,8 @@ with gr.Blocks(css=css) as demo:
|
|
246 |
with gr.Row():
|
247 |
with gr.Column():
|
248 |
chatbot = gr.Chatbot(elem_id="chatbox")
|
249 |
-
input_message = gr.Textbox(show_label=False, placeholder=docs, visible=True).style(container=False)
|
250 |
-
|
251 |
|
252 |
btn_submit = gr.Button("Submit")
|
253 |
total_tokens_str = gr.Markdown(elem_id="total_tokens_str")
|
|
|
104 |
#vectordb = Chroma.from_documents(split_pages, embeddings, persist_directory=persist_directory)
|
105 |
#vectordb.persist()
|
106 |
|
107 |
+
#path = './files'
|
108 |
+
#pdf_files = glob.glob(os.path.join(path, "*.pdf"))
|
109 |
|
110 |
+
#merger = PdfWriter()
|
111 |
|
112 |
# add all file in the list to the merger object
|
113 |
+
#for pdf in pdf_files:
|
114 |
+
# merger.append(pdf)
|
115 |
+
#merger.write("merged-pdf.pdf")
|
116 |
+
#merger.close()
|
117 |
|
118 |
+
#reader = PdfReader("merged-pdf.pdf")
|
119 |
+
#raw_text = ''
|
120 |
+
#for i, page in enumerate(reader.pages):
|
121 |
+
# text = page.extract_text()
|
122 |
+
# if text:
|
123 |
+
# raw_text += text
|
124 |
+
#text_splitter = CharacterTextSplitter(
|
125 |
+
# separator = "\n",
|
126 |
+
# chunk_size = 1000,
|
127 |
+
# chunk_overlap = 200,
|
128 |
+
# length_function = len,
|
129 |
+
#)
|
130 |
+
#texts = text_splitter.split_text(raw_text)
|
131 |
+
#len(texts)
|
132 |
+
#embeddings = OpenAIEmbeddings()
|
133 |
|
134 |
|
135 |
history = state['messages']
|
|
|
168 |
#with open("foo.pkl", 'rb') as f:
|
169 |
# new_docsearch = pickle.load(f)
|
170 |
|
171 |
+
#docsearch = FAISS.from_texts(texts, embeddings)
|
172 |
+
with open("embeddings.pkl", 'rb') as f:
|
173 |
+
new_docsearch = pickle.load(f)
|
174 |
+
|
175 |
#query = str(system_prompt + history[-context_length*2:] + [prompt_msg])
|
176 |
query = str(system_prompt + history + [prompt_msg])
|
177 |
+
docs = new_docsearch.similarity_search(query)
|
178 |
#print(docs[0].page_content)
|
179 |
|
180 |
chain = load_qa_chain(ChatOpenAI(temperature=temperature, max_tokens=max_tokens, model_name="gpt-3.5-turbo"), chain_type="stuff")
|
|
|
232 |
|
233 |
|
234 |
with gr.Column(elem_id="col-container"):
|
235 |
+
#with open("embeddings.pkl", 'rb') as f:
|
236 |
+
# new_docsearch = pickle.load(f)
|
237 |
|
238 |
+
#query = str("performance")
|
239 |
+
#docs = new_docsearch.similarity_search(query)
|
240 |
|
241 |
gr.Markdown("""# Chat with Needs Assessment Experts (Past and Present)
|
242 |
## Ask questions of experts on needs assessments, get responses from *needs assessment* version of ChatGPT.
|
|
|
249 |
with gr.Row():
|
250 |
with gr.Column():
|
251 |
chatbot = gr.Chatbot(elem_id="chatbox")
|
252 |
+
#input_message = gr.Textbox(show_label=False, placeholder=docs, visible=True).style(container=False)
|
253 |
+
input_message = gr.Textbox(show_label=False, placeholder="Enter your needs assessment question and press enter", visible=True).style(container=False)
|
254 |
|
255 |
btn_submit = gr.Button("Submit")
|
256 |
total_tokens_str = gr.Markdown(elem_id="total_tokens_str")
|