Spaces:
Sleeping
Sleeping
Ayush Mangal
commited on
Commit
β’
7610853
1
Parent(s):
fe10a2c
Remove adding rep key
Browse files
app.py
CHANGED
@@ -11,10 +11,15 @@ from langchain.embeddings import HuggingFaceEmbeddings
|
|
11 |
from langchain.chains import RetrievalQA
|
12 |
from langchain.vectorstores.base import VectorStoreRetriever
|
13 |
from langchain.prompts import PromptTemplate
|
|
|
|
|
|
|
14 |
|
15 |
-
st.set_page_config(page_title="π¦π
|
16 |
-
st.title('π¦π
|
17 |
-
|
|
|
|
|
18 |
|
19 |
@st.cache_resource
|
20 |
def get_query_chain():
|
@@ -47,6 +52,23 @@ def get_query_chain():
|
|
47 |
print("got loader")
|
48 |
data = loader.load()
|
49 |
video_data.extend(data)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
50 |
text_splitter = RecursiveCharacterTextSplitter(chunk_size = 100, chunk_overlap = 0)
|
51 |
all_splits = text_splitter.split_documents(video_data)
|
52 |
vectorstore = FAISS.from_documents(documents=all_splits, embedding=hf)
|
@@ -81,6 +103,12 @@ with st.form('myform'):
|
|
81 |
topic_text = st.text_input('Enter keyword:', '')
|
82 |
submitted = st.form_submit_button('Submit')
|
83 |
if submitted :
|
84 |
-
os.environ["REPLICATE_API_TOKEN"] =
|
85 |
query_chain = get_query_chain()
|
86 |
-
generate_response(topic_text, query_chain)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
from langchain.chains import RetrievalQA
|
12 |
from langchain.vectorstores.base import VectorStoreRetriever
|
13 |
from langchain.prompts import PromptTemplate
|
14 |
+
from lxml import etree
|
15 |
+
from langchain.document_loaders import WebBaseLoader
|
16 |
+
from bs4 import BeautifulSoup
|
17 |
|
18 |
+
st.set_page_config(page_title="π¦π Ask an LLM to know more about me")
|
19 |
+
st.title('π¦π Ask an LLM to know more about me')
|
20 |
+
|
21 |
+
def has_numbers(inputString):
|
22 |
+
return any(char.isdigit() for char in inputString)
|
23 |
|
24 |
@st.cache_resource
|
25 |
def get_query_chain():
|
|
|
52 |
print("got loader")
|
53 |
data = loader.load()
|
54 |
video_data.extend(data)
|
55 |
+
profile_url = "https://ayushtues.medium.com"
|
56 |
+
response = requests.get(profile_url)
|
57 |
+
soup = BeautifulSoup(response.content, 'html.parser')
|
58 |
+
links = []
|
59 |
+
for link in soup.findAll('a'):
|
60 |
+
x = link.get('href')
|
61 |
+
if x.startswith('/') and has_numbers(x) :
|
62 |
+
links.append(link.get('href'))
|
63 |
+
links = list(set(links))
|
64 |
+
|
65 |
+
links = [profile_url+ x.split('?source')[0] for x in links]
|
66 |
+
# print(links)
|
67 |
+
|
68 |
+
loader = WebBaseLoader(links)
|
69 |
+
data = loader.load()
|
70 |
+
video_data.extend(data)
|
71 |
+
# print(data)
|
72 |
text_splitter = RecursiveCharacterTextSplitter(chunk_size = 100, chunk_overlap = 0)
|
73 |
all_splits = text_splitter.split_documents(video_data)
|
74 |
vectorstore = FAISS.from_documents(documents=all_splits, embedding=hf)
|
|
|
103 |
topic_text = st.text_input('Enter keyword:', '')
|
104 |
submitted = st.form_submit_button('Submit')
|
105 |
if submitted :
|
106 |
+
os.environ["REPLICATE_API_TOKEN"] = st.secrets["REPLICATE_API_TOKEN"]
|
107 |
query_chain = get_query_chain()
|
108 |
+
generate_response(topic_text, query_chain)
|
109 |
+
|
110 |
+
|
111 |
+
|
112 |
+
|
113 |
+
|
114 |
+
|