from langchain_openai import ChatOpenAI import os try: from dotenv import load_dotenv load_dotenv() except Exception: pass # gpt-3.5-turbo-0125 def get_llm(model="gpt-3.5-turbo", max_tokens=1024, temperature=0.0, streaming=True, timeout=30, **kwargs): llm = ChatOpenAI( model=model, api_key=os.environ.get("OPENAI_API_KEY", None), max_tokens=max_tokens, streaming=streaming, temperature=temperature, timeout=timeout, **kwargs, ) return llm