AI4PE-STATIC / hogwarts.py
Adr740's picture
Upload 4 files
f798440 verified
import numpy as np
from openai import OpenAI
from groq import Groq
import pandas as pd
from config import openai_api, groq_api, models
provider = "openai"
if provider == "openai":
client = OpenAI(api_key=openai_api)
else:
client = Groq(api_key=groq_api)
system_prompt_default = """
"""
def cosine_similarity(a, b):
return np.dot(a, b) / (np.linalg.norm(a) * np.linalg.norm(b))
def _get_embedding(text, model="text-embedding-3-large"):
try:
text = text.replace("\n", " ")
except:
None
return client.embeddings.create(input = [text], model=model).data[0].embedding
def get_answer(df, nb_in_context = 10, task = "Your task is to estimate the revenue evolution of the considered company."):
# query = str(query_preprocessor_augment(task))
embedding_query = _get_embedding(task, model="text-embedding-3-large")
try:
df['similarity'] = df.embeddings.apply(lambda x: cosine_similarity(x, embedding_query))
except:
df['similarity'] = df.embeddings.apply(lambda x: cosine_similarity(eval(x), embedding_query))
res = df.sort_values('similarity', ascending=False).head(nb_in_context).content.values
response = client.chat.completions.create(
model="gpt-4o",
messages=[
{
"role": "system",
"content": [
{
"type": "text",
"text": f"""You will be given a vast amount of data, each with it's source. {task}
Your answer, should be crisp, sharp and pinpoint to an exact source from the context (if the references of the sources are not easy to read by humans feel free to adjust so that it's readable - no links though, you refer to them as what they are).
You write using Bain's style as it will be read by private equity professionals and if asked, you never refer to yourself.
"""
}
]
},
{
"role": "user",
"content": [
{
"type": "text",
"text": f"Context:\n{str(res)}"
}
]
}
],
temperature=1,
max_tokens=1665,
top_p=1,
frequency_penalty=0,
presence_penalty=0
).choices[0].message.content
return response
# for r in res:
# print(r)
def generate_content(input, data_dumpster, chunked_raw_content, custom_query = (False, "")):
data_locker_folder = f"./{data_dumpster}/{input.replace(' ','-')}"
try:
df = pd.read_csv(f"{data_locker_folder}/vectorized_data_dumpster.csv", sep=";")
except:
df = pd.DataFrame()
df["content"] = chunked_raw_content
embeddings = []
for chunk in chunked_raw_content:
embeddings.append(_get_embedding(chunk))
df["embeddings"] = embeddings
df.to_csv(f"{data_locker_folder}/vectorized_data_dumpster.csv", sep=";")
finance = "Your task is to estimate the revenue evolution of the considered company."
product = "Your task is to give an overview of the line of products of the considered company."
customer = "Your task is to give the probable customer segmentation of the considered company."
in_context = 15
if custom_query[0]:
print("Generating custom chat output")
custom_answer = get_answer(df, in_context, custom_query[1])
return custom_answer
print("Generating financials content")
finance_content = get_answer(df, in_context, finance)
print("Generating product content")
product_content = get_answer(df, in_context, product)
print("Generating customer segmentation content")
customer_content = get_answer(df, in_context, customer)
print("Done!")
rag_content = {
"finance" : finance_content,
"product": product_content,
"customer_segmentation": customer_content
}
return rag_content