chat-with-your-pdf / utils.py
anilkumar-kanasani's picture
Upload 3 files
e6f8d33
from environs import Env
env = Env()
try:
env.read_env("/Users/kanasani/Documents/api_keys/.env.llm")
print("Using local .env.llm file")
except:
env.read_env()
print(".env file from repo secrets is used")
import openai
openai.api_type = env("API_TYPE")
openai.api_base = env("API_BASE")
openai.api_version = env("API_VERSION")
openai.api_key = env("AZURE_OPENAI_KEY")
def check_password():
import streamlit as st
"""Returns `True` if the user had the correct password."""
def password_entered():
"""Checks whether a password entered by the user is correct."""
if st.session_state["password"] == env("st_password"):
st.session_state["password_correct"] = True
del st.session_state["password"] # don't store password
else:
st.session_state["password_correct"] = False
if "password_correct" not in st.session_state:
# First run, show input for password.
st.text_input(
"Password", type="password", on_change=password_entered, key="password"
)
return False
elif not st.session_state["password_correct"]:
# Password not correct, show input + error.
st.text_input(
"Password", type="password", on_change=password_entered, key="password"
)
st.error("πŸ˜• Password incorrect")
return False
else:
# Password correct.
return True
def submit_prompt_to_gpt(input_list_of_prompts):
response = openai.ChatCompletion.create(
engine=env("DEPLOYMENT_NAME"),
messages=input_list_of_prompts,
temperature=1,
max_tokens=256,
top_p=1,
frequency_penalty=0,
presence_penalty=0,
)
response_content = response["choices"][0]["message"]["content"]
return response_content
def get_hf_embeddings():
from langchain.embeddings import HuggingFaceHubEmbeddings
embeddings = HuggingFaceHubEmbeddings(
repo_id="sentence-transformers/all-mpnet-base-v2",
task="feature-extraction",
huggingfacehub_api_token=env("HUGGINGFACEHUB_API_TOKEN"),
)
return embeddings
def get_openAI_chat_model():
import openai
from langchain.chat_models.azure_openai import AzureChatOpenAI
chat_model = AzureChatOpenAI(deployment_name=env("DEPLOYMENT_NAME"),
openai_api_version=env("API_VERSION"),
openai_api_base=env("API_BASE"),
openai_api_type=env("API_TYPE"),
openai_api_key=env("AZURE_OPENAI_KEY"),
verbose=True)
return chat_model
def get_hf_model(repo_id = "google/flan-t5-xxl"):
from langchain import HuggingFaceHub
hf_llm = HuggingFaceHub(
repo_id=repo_id,
model_kwargs={"temperature": 0.1, "max_length": 1024},
huggingfacehub_api_token = env("HUGGINGFACEHUB_API_TOKEN"),
)
return hf_llm
def get_local_gpt4_model(model = "GPT4All-13B-snoozy.ggmlv3.q4_0.bin"):
from langchain.llms import GPT4All
gpt4_llm = GPT4All(model=".models/"+model,
verbose=True)
return gpt4_llm
def set_LangChain_tracking(project="Chat with your PDF"):
import os
os.environ['LANGCHAIN_PROJECT'] = project
print("LangChain tracking is set to : ", project)
def unset_LangChain_tracking():
import os
os.environ.pop('LANGCHAIN_API_KEY', None)
os.environ.pop('LANGCHAIN_TRACING_V2', None)
os.environ.pop('LANGCHAIN_ENDPOINT', None)
os.environ.pop('LANGCHAIN_PROJECT', None)
print("LangChain tracking is removed .")