import torch.cuda
import torch.backends


prompt_model_dict = {
    "lexart": "/home/chia/model/AUTOMATIC_promptgen-lexart"
}

# Prompt generate model name
PROMPT_MODEL = "lexart"

# Embedding running device
PROMPT_DEVICE = "cuda" if torch.cuda.is_available() else "mps" if torch.backends.mps.is_available() else "cpu"


# supported LLM models
llm_model_dict = {
    "alpaca": "/home/chia/model/tatsu_alpaca-7b",
    "llama": "/home/chia/model/Meta_LLaMA_hf/llama-7b-hf"
}

# LLM model name
LLM_MODEL = "alpaca"

# LLM running device
LLM_DEVICE = "cuda" if torch.cuda.is_available() else "mps" if torch.backends.mps.is_available() else "cpu"
