Spaces:
Runtime error
Runtime error
from langchain.chat_models import ChatOpenAI | |
def load_llama_model(): | |
return ChatOpenAI( | |
#Llama 3 8B 모델로 RAG 실행하고 싶은 경우 | |
#base_url="http://torch27:8000/v1", | |
#model="meta-llama/Meta-Llama-3-8B-Instruct", | |
#openai_api_key="EMPTY" | |
#Exaone으로 RAG 실행하고 싶은 경우 | |
base_url="http://220.124.155.35:8000/v1", | |
model="LGAI-EXAONE/EXAONE-3.5-7.8B-Instruct", | |
openai_api_key="token-abc123" | |
#base_url="https://7xiebe4unotxnp-8000.proxy.runpod.net/v1", | |
#model="meta-llama/Llama-4-Scout-17B-16E-Instruct", | |
#openai_api_key="EMPTY" | |
# base_url="http://vllm_yjy:8000/v1", | |
# model="/models/Llama-3.3-70B-Instruct-AWQ", | |
# openai_api_key="token-abc123" | |
) | |