sameemul-haque commited on
Commit
f7f0473
1 Parent(s): ae244b4

wip: flask

Browse files
Files changed (2) hide show
  1. .gitignore +2 -1
  2. app.py +22 -8
.gitignore CHANGED
@@ -1,4 +1,5 @@
1
  Documents
2
  .env
3
  venv
4
- test
 
 
1
  Documents
2
  .env
3
  venv
4
+ test
5
+ __pycache__
app.py CHANGED
@@ -8,8 +8,16 @@ from langchain_community.document_loaders import PyPDFLoader
8
  from langchain_community.document_loaders import DirectoryLoader
9
  from langchain.text_splitter import RecursiveCharacterTextSplitter
10
  from langchain_community.embeddings import HuggingFaceInstructEmbeddings
 
 
 
 
 
11
 
12
  def main():
 
 
 
13
  # load env
14
  load_dotenv()
15
 
@@ -26,7 +34,10 @@ def main():
26
  # create the retriever
27
  db_instructEmbedd = FAISS.from_documents(texts, instructor_embeddings)
28
  retriever = db_instructEmbedd.as_retriever(search_kwargs={"k": 3})
29
- query = 'What is operating system?'
 
 
 
30
  # print('retriever search type:',retriever.search_type) # retriever search type is similarity search
31
  # print('retriever search kwargs:',retriever.search_kwargs)
32
  # docs = retriever.get_relevant_documents(query)
@@ -50,15 +61,18 @@ def main():
50
  # Join the wrapped lines back together using newline characters
51
  wrapped_text = '\n'.join(wrapped_lines)
52
  return wrapped_text
53
-
54
- def process_llm_response(llm_response):
55
- print(wrap_text_preserve_newlines(llm_response['result']))
56
- print('\nSources:')
57
- for source in llm_response["source_documents"]:
58
- print(source.metadata['source'])
59
 
60
  llm_response = qa_chain_instrucEmbed(query)
61
- process_llm_response(llm_response)
 
 
 
 
62
 
63
  if __name__ == '__main__':
64
  main()
 
8
  from langchain_community.document_loaders import DirectoryLoader
9
  from langchain.text_splitter import RecursiveCharacterTextSplitter
10
  from langchain_community.embeddings import HuggingFaceInstructEmbeddings
11
+ from flask import Flask, request
12
+
13
+ app = Flask(__name__)
14
+
15
+ @app.route('/',methods=['GET'])
16
 
17
  def main():
18
+ query = request.args.get('q')
19
+ # query = unquote(query)
20
+
21
  # load env
22
  load_dotenv()
23
 
 
34
  # create the retriever
35
  db_instructEmbedd = FAISS.from_documents(texts, instructor_embeddings)
36
  retriever = db_instructEmbedd.as_retriever(search_kwargs={"k": 3})
37
+ # print("this is the embeddings----------:",db_instructEmbedd,"----------embeddings")
38
+ # print("this is the retriever----------:",retriever,"----------retriever")
39
+ # db.save_local("faiss_index")
40
+ # query = 'What is operating system?'
41
  # print('retriever search type:',retriever.search_type) # retriever search type is similarity search
42
  # print('retriever search kwargs:',retriever.search_kwargs)
43
  # docs = retriever.get_relevant_documents(query)
 
61
  # Join the wrapped lines back together using newline characters
62
  wrapped_text = '\n'.join(wrapped_lines)
63
  return wrapped_text
64
+ # def process_llm_response(llm_response):
65
+ # print(wrap_text_preserve_newlines(llm_response['result']))
66
+ # print('\nSources:')
67
+ # for source in llm_response["source_documents"]:
68
+ # print(source.metadata['source'])
 
69
 
70
  llm_response = qa_chain_instrucEmbed(query)
71
+ res = wrap_text_preserve_newlines(llm_response['result'])
72
+ # process_llm_response(llm_response)
73
+ return res
74
+
75
+ #print(res,"result")
76
 
77
  if __name__ == '__main__':
78
  main()