Update chat_with_doc.py
Browse files- chat_with_doc.py +3 -5
chat_with_doc.py
CHANGED
@@ -1,7 +1,6 @@
|
|
1 |
import os
|
2 |
import openai
|
3 |
import streamlit as st
|
4 |
-
import json
|
5 |
|
6 |
# Load environment variables
|
7 |
from dotenv import load_dotenv
|
@@ -25,10 +24,10 @@ def get_openai_response(prompt, model="gpt-4", max_tokens=150):
|
|
25 |
# Streamlit UI
|
26 |
st.title("Chat with Your Document")
|
27 |
|
28 |
-
# Load chunked data from the file (chunked_data.
|
29 |
try:
|
30 |
-
with open("chunked_data.
|
31 |
-
chunked_data =
|
32 |
st.write("Document has been chunked into the following parts:")
|
33 |
for i, chunk_part in enumerate(chunked_data, 1):
|
34 |
st.write(f"**Chunk {i}:**\n{chunk_part}\n")
|
@@ -48,4 +47,3 @@ if st.button("Get Answer"):
|
|
48 |
st.write(f"**Answer**: {response}")
|
49 |
else:
|
50 |
st.error("Please provide a question.")
|
51 |
-
|
|
|
1 |
import os
|
2 |
import openai
|
3 |
import streamlit as st
|
|
|
4 |
|
5 |
# Load environment variables
|
6 |
from dotenv import load_dotenv
|
|
|
24 |
# Streamlit UI
|
25 |
st.title("Chat with Your Document")
|
26 |
|
27 |
+
# Load chunked data from the .txt file (chunked_data.txt)
|
28 |
try:
|
29 |
+
with open("chunked_data.txt", "r") as f:
|
30 |
+
chunked_data = f.read().split("\n---\n") # Split by delimiter to separate chunks
|
31 |
st.write("Document has been chunked into the following parts:")
|
32 |
for i, chunk_part in enumerate(chunked_data, 1):
|
33 |
st.write(f"**Chunk {i}:**\n{chunk_part}\n")
|
|
|
47 |
st.write(f"**Answer**: {response}")
|
48 |
else:
|
49 |
st.error("Please provide a question.")
|
|