litellm / litellm_config.yaml
rkihacker's picture
Update litellm_config.yaml
2e888e9 verified
# LiteLLM Proxy config for Replicate
# Save as config.yaml and run: litellm --config ./config.yaml --debug
model_list:
# Meta Llama 3 Instruct (8B) via Replicate
- model_name: openai/gpt-5-nano
litellm_params:
model: replicate/openai/gpt-5-nano
api_key: os.environ/REPLICATE_API_KEY
# optional common params
temperature: 0.7
max_tokens: 1024
# Llama-2-70B chat with a pinned version ID (replace the example hash with the current one you use)
- model_name: openai/gpt-5-mini
litellm_params:
model: replicate/openai/gpt-5-mini
api_key: os.environ/REPLICATE_API_KEY
# Call a Replicate *deployment* you own (replace owner/name)
- model_name: openai/gpt-5
litellm_params:
model: replicate/openai/gpt-5
api_key: os.environ/REPLICATE_API_KEY
# Optional: proxy/router tuning
router_settings:
timeout: 600 # Replicate can cold-boot, so give it time
num_retries: 1
# (Optional) enable simple logging callbacks, etc.
# litellm_settings:
# success_callback: ["generic"]