litellm_settings: drop_params: True # Model-specific settings model_list: # use the same model_name for using the litellm router. LiteLLM will use the router between gpt-3.5-turbo - model_name: gpt-3.5-turbo # litellm will litellm_params: model: gpt-3.5-turbo api_key: sk-uj6F tpm: 20000 # [OPTIONAL] REPLACE with your openai tpm rpm: 3 # [OPTIONAL] REPLACE with your openai rpm - model_name: gpt-3.5-turbo litellm_params: model: gpt-3.5-turbo api_key: sk-Imn tpm: 20000 # [OPTIONAL] REPLACE with your openai tpm rpm: 3 # [OPTIONAL] REPLACE with your openai rpm - model_name: gpt-3.5-turbo litellm_params: model: openrouter/gpt-3.5-turbo - model_name: mistral-7b-instruct litellm_params: model: mistralai/mistral-7b-instruct environment_variables: REDIS_HOST: localhost REDIS_PASSWORD: REDIS_PORT: