Ollama-RAG / utils /QnA.py
Mr-TD's picture
Test
9dc99bc
raw
history blame contribute delete
359 Bytes
from langchain.chains import RetrievalQA
from langchain_ollama import ChatOllama
def Q_A(vectorstore,question):
ollama_llm = ChatOllama(
model="llama3.1",
temperature=0.5,
)
qa = RetrievalQA.from_chain_type(llm=ollama_llm, chain_type="stuff", retriever=vectorstore.as_retriever())
answer = qa.invoke(question)
return answer['result']