Spaces:
Runtime error
Runtime error
testing with llama3
Browse files- src/app.py +2 -2
- src/interface.py +1 -1
- src/pdfchatbot.py +2 -4
src/app.py
CHANGED
@@ -2,7 +2,7 @@ from interface import create_demo
|
|
2 |
from pdfchatbot import PDFChatBot
|
3 |
|
4 |
# Create Gradio interface
|
5 |
-
demo, chat_history, show_img, txt, submit_button, uploaded_pdf,
|
6 |
|
7 |
# Create PDFChatBot instance
|
8 |
pdf_chatbot = PDFChatBot()
|
@@ -12,7 +12,7 @@ pdf_chatbot.load_tokenizer()
|
|
12 |
# Set up event handlers
|
13 |
with demo:
|
14 |
# Event handler for uploading a PDF
|
15 |
-
uploaded_pdf.upload(pdf_chatbot.render_file, inputs=[uploaded_pdf
|
16 |
|
17 |
# Event handler for submitting text and generating response
|
18 |
submit_button.click(pdf_chatbot.add_text, inputs=[chat_history, txt], outputs=[chat_history], queue=False).\
|
|
|
2 |
from pdfchatbot import PDFChatBot
|
3 |
|
4 |
# Create Gradio interface
|
5 |
+
demo, chat_history, show_img, txt, submit_button, uploaded_pdf, _ = create_demo()
|
6 |
|
7 |
# Create PDFChatBot instance
|
8 |
pdf_chatbot = PDFChatBot()
|
|
|
12 |
# Set up event handlers
|
13 |
with demo:
|
14 |
# Event handler for uploading a PDF
|
15 |
+
uploaded_pdf.upload(pdf_chatbot.render_file, inputs=[uploaded_pdf], outputs=[show_img])
|
16 |
|
17 |
# Event handler for submitting text and generating response
|
18 |
submit_button.click(pdf_chatbot.add_text, inputs=[chat_history, txt], outputs=[chat_history], queue=False).\
|
src/interface.py
CHANGED
@@ -11,7 +11,7 @@ def create_demo():
|
|
11 |
with gr.Row():
|
12 |
# Add sliders here
|
13 |
with gr.Column(): # Adjust scale as needed
|
14 |
-
slider1 = gr.Slider(minimum=256, maximum=1024, value=
|
15 |
with gr.Row():
|
16 |
with gr.Column(scale=0.60):
|
17 |
text_input = gr.Textbox(
|
|
|
11 |
with gr.Row():
|
12 |
# Add sliders here
|
13 |
with gr.Column(): # Adjust scale as needed
|
14 |
+
slider1 = gr.Slider(minimum=256, maximum=1024, value=256, label="Chunk Size")
|
15 |
with gr.Row():
|
16 |
with gr.Column(scale=0.60):
|
17 |
text_input = gr.Textbox(
|
src/pdfchatbot.py
CHANGED
@@ -95,7 +95,7 @@ class PDFChatBot:
|
|
95 |
"""
|
96 |
Load the vector database from the documents and embeddings.
|
97 |
"""
|
98 |
-
text_splitter = CharacterTextSplitter(chunk_size=self.chunk_size, chunk_overlap=
|
99 |
docs = text_splitter.split_documents(self.documents)
|
100 |
self.vectordb = Chroma.from_documents(docs, self.embeddings)
|
101 |
|
@@ -178,13 +178,11 @@ class PDFChatBot:
|
|
178 |
|
179 |
result = self.chain({"question": query, 'chat_history': self.chat_history}, return_only_outputs=True)
|
180 |
self.chat_history.append((query, result["answer"]))
|
181 |
-
self.page = list(result['source_documents'][0])[1][1]['page']
|
182 |
-
|
183 |
for char in result['answer']:
|
184 |
history[-1][-1] += char
|
185 |
return history, " "
|
186 |
|
187 |
-
def render_file(self, file,chunk_size):
|
188 |
"""
|
189 |
Renders a specific page of a PDF file as an image.
|
190 |
|
|
|
95 |
"""
|
96 |
Load the vector database from the documents and embeddings.
|
97 |
"""
|
98 |
+
text_splitter = CharacterTextSplitter(chunk_size=self.chunk_size, chunk_overlap=100)
|
99 |
docs = text_splitter.split_documents(self.documents)
|
100 |
self.vectordb = Chroma.from_documents(docs, self.embeddings)
|
101 |
|
|
|
178 |
|
179 |
result = self.chain({"question": query, 'chat_history': self.chat_history}, return_only_outputs=True)
|
180 |
self.chat_history.append((query, result["answer"]))
|
|
|
|
|
181 |
for char in result['answer']:
|
182 |
history[-1][-1] += char
|
183 |
return history, " "
|
184 |
|
185 |
+
def render_file(self, file,chunk_size=256):
|
186 |
"""
|
187 |
Renders a specific page of a PDF file as an image.
|
188 |
|