Spaces:
Sleeping
Sleeping
| import os | |
| import numpy as np | |
| from langchain_groq import ChatGroq | |
| from langchain_openai import ChatOpenAI | |
| from langchain_google_genai import ChatGoogleGenerativeAI | |
| from langchain_anthropic import ChatAnthropic | |
| from langchain_openai import OpenAIEmbeddings | |
| from huggingface_hub import login | |
| from dotenv import load_dotenv | |
| import tiktoken | |
| load_dotenv() | |
| os.environ['HUGGINGFACEHUB_ACCESS_TOKEN']=os.getenv('HUGGINGFACEHUB_ACCESS_TOKEN') | |
| login(os.environ['HUGGINGFACEHUB_ACCESS_TOKEN']) | |
| embedding_model = OpenAIEmbeddings(model="text-embedding-3-small", dimensions=1536) | |
| llm_anthropic = ChatAnthropic(model='claude-3-7-sonnet-latest', temperature=1) | |
| llm_gemini = ChatGoogleGenerativeAI(model="gemini-1.5-flash") | |
| llm_groq_openai = ChatGroq(model="openai/gpt-oss-120b",temperature=0.7) | |
| llm_groq = ChatGroq(model="llama-3.3-70b-versatile",temperature=0) | |
| llm_gpt_small = ChatOpenAI(model="gpt-3.5-turbo",temperature=0.3) | |
| llm_gpt = ChatOpenAI(model="gpt-3.5-turbo",temperature=0.3) | |
| llm_gpt_high = ChatOpenAI(model="gpt-5-nano",temperature=0.5) | |
| # encoding_model = tiktoken.encoding_for_model('gpt-4o-mini') | |
| encoding_model = 'encoding_model' | |
| captioning_model = "meta-llama/llama-4-scout-17b-16e-instruct" | |
| image_generation_model = "black-forest-labs/FLUX.1-schnell" | |
| ideator_llm = llm_gpt_high | |
| moderator_llm = llm_gpt | |
| critic_llm = llm_gpt | |
| simplifier_llm = llm_gpt | |
| normalizer_llm = llm_gpt | |
| validator_llm = llm_gpt_small | |
| judge1_llm = llm_gpt | |
| judge2_llm = llm_gpt | |