Ayush Mangal commited on
Commit
7610853
β€’
1 Parent(s): fe10a2c

Remove adding rep key

Browse files
Files changed (1) hide show
  1. app.py +33 -5
app.py CHANGED
@@ -11,10 +11,15 @@ from langchain.embeddings import HuggingFaceEmbeddings
11
  from langchain.chains import RetrievalQA
12
  from langchain.vectorstores.base import VectorStoreRetriever
13
  from langchain.prompts import PromptTemplate
 
 
 
14
 
15
- st.set_page_config(page_title="πŸ¦œπŸ”— YT Summarizer")
16
- st.title('πŸ¦œπŸ”— YT Summarizer')
17
- jina_api_key = st.sidebar.text_input('Replicate API Key', type='password')
 
 
18
 
19
  @st.cache_resource
20
  def get_query_chain():
@@ -47,6 +52,23 @@ def get_query_chain():
47
  print("got loader")
48
  data = loader.load()
49
  video_data.extend(data)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
50
  text_splitter = RecursiveCharacterTextSplitter(chunk_size = 100, chunk_overlap = 0)
51
  all_splits = text_splitter.split_documents(video_data)
52
  vectorstore = FAISS.from_documents(documents=all_splits, embedding=hf)
@@ -81,6 +103,12 @@ with st.form('myform'):
81
  topic_text = st.text_input('Enter keyword:', '')
82
  submitted = st.form_submit_button('Submit')
83
  if submitted :
84
- os.environ["REPLICATE_API_TOKEN"] = jina_api_key
85
  query_chain = get_query_chain()
86
- generate_response(topic_text, query_chain)
 
 
 
 
 
 
 
11
  from langchain.chains import RetrievalQA
12
  from langchain.vectorstores.base import VectorStoreRetriever
13
  from langchain.prompts import PromptTemplate
14
+ from lxml import etree
15
+ from langchain.document_loaders import WebBaseLoader
16
+ from bs4 import BeautifulSoup
17
 
18
+ st.set_page_config(page_title="πŸ¦œπŸ”— Ask an LLM to know more about me")
19
+ st.title('πŸ¦œπŸ”— Ask an LLM to know more about me')
20
+
21
+ def has_numbers(inputString):
22
+ return any(char.isdigit() for char in inputString)
23
 
24
  @st.cache_resource
25
  def get_query_chain():
 
52
  print("got loader")
53
  data = loader.load()
54
  video_data.extend(data)
55
+ profile_url = "https://ayushtues.medium.com"
56
+ response = requests.get(profile_url)
57
+ soup = BeautifulSoup(response.content, 'html.parser')
58
+ links = []
59
+ for link in soup.findAll('a'):
60
+ x = link.get('href')
61
+ if x.startswith('/') and has_numbers(x) :
62
+ links.append(link.get('href'))
63
+ links = list(set(links))
64
+
65
+ links = [profile_url+ x.split('?source')[0] for x in links]
66
+ # print(links)
67
+
68
+ loader = WebBaseLoader(links)
69
+ data = loader.load()
70
+ video_data.extend(data)
71
+ # print(data)
72
  text_splitter = RecursiveCharacterTextSplitter(chunk_size = 100, chunk_overlap = 0)
73
  all_splits = text_splitter.split_documents(video_data)
74
  vectorstore = FAISS.from_documents(documents=all_splits, embedding=hf)
 
103
  topic_text = st.text_input('Enter keyword:', '')
104
  submitted = st.form_submit_button('Submit')
105
  if submitted :
106
+ os.environ["REPLICATE_API_TOKEN"] = st.secrets["REPLICATE_API_TOKEN"]
107
  query_chain = get_query_chain()
108
+ generate_response(topic_text, query_chain)
109
+
110
+
111
+
112
+
113
+
114
+