Spaces:
Runtime error
Runtime error
| from flask import Flask, render_template, jsonify, request | |
| from src.helper import download_hugging_face_embeddings | |
| from langchain.vectorstores import Pinecone | |
| import pinecone | |
| from langchain.prompts import PromptTemplate | |
| from langchain.llms import CTransformers | |
| from langchain.chains import RetrievalQA | |
| from dotenv import load_dotenv | |
| from src.prompt import * | |
| import os | |
| app = Flask(__name__) | |
| load_dotenv() | |
| PINECONE_API_KEY = os.environ.get('PINECONE_API_KEY') | |
| PINECONE_API_ENV = os.environ.get('PINECONE_API_ENV') | |
| embeddings = download_hugging_face_embeddings() | |
| #Initializing the Pinecone | |
| pinecone.init(api_key=PINECONE_API_KEY, | |
| environment=PINECONE_API_ENV) | |
| index_name="geeta-bot" | |
| #Loading the index | |
| docsearch=Pinecone.from_existing_index(index_name, embeddings) | |
| PROMPT=PromptTemplate(template=prompt_template, input_variables=["context", "question"]) | |
| chain_type_kwargs={"prompt": PROMPT} | |
| llm=CTransformers(model="model/llama-2-7b-chat.ggmlv3.q4_0.bin", | |
| model_type="llama", | |
| config={'max_new_tokens':512, | |
| 'temperature':0.8}) | |
| qa=RetrievalQA.from_chain_type( | |
| llm=llm, | |
| chain_type="stuff", | |
| retriever=docsearch.as_retriever(search_kwargs={'k': 2}), | |
| return_source_documents=True, | |
| chain_type_kwargs=chain_type_kwargs) | |
| def index(): | |
| return render_template('chat.html') | |
| def chat(): | |
| msg = request.form["msg"] | |
| input = msg | |
| print(input) | |
| result=qa({"query": input}) | |
| print("Response : ", result["result"]) | |
| return str(result["result"]) | |
| if __name__ == '__main__': | |
| app.run(host="0.0.0.0", port= 7860, debug= True) | |