|
{ |
|
"gpt-4": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.00003, |
|
"output_cost_per_token": 0.00006, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"gpt-4-0314": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.00003, |
|
"output_cost_per_token": 0.00006, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"gpt-4-0613": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.00003, |
|
"output_cost_per_token": 0.00006, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"gpt-4-32k": { |
|
"max_tokens": 32768, |
|
"input_cost_per_token": 0.00006, |
|
"output_cost_per_token": 0.00012, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"gpt-4-32k-0314": { |
|
"max_tokens": 32768, |
|
"input_cost_per_token": 0.00006, |
|
"output_cost_per_token": 0.00012, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"gpt-4-32k-0613": { |
|
"max_tokens": 32768, |
|
"input_cost_per_token": 0.00006, |
|
"output_cost_per_token": 0.00012, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"gpt-3.5-turbo": { |
|
"max_tokens": 4097, |
|
"input_cost_per_token": 0.0000015, |
|
"output_cost_per_token": 0.000002, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"gpt-3.5-turbo-0301": { |
|
"max_tokens": 4097, |
|
"input_cost_per_token": 0.0000015, |
|
"output_cost_per_token": 0.000002, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"gpt-3.5-turbo-0613": { |
|
"max_tokens": 4097, |
|
"input_cost_per_token": 0.0000015, |
|
"output_cost_per_token": 0.000002, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"gpt-3.5-turbo-16k": { |
|
"max_tokens": 16385, |
|
"input_cost_per_token": 0.000003, |
|
"output_cost_per_token": 0.000004, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"gpt-3.5-turbo-16k-0613": { |
|
"max_tokens": 16385, |
|
"input_cost_per_token": 0.000003, |
|
"output_cost_per_token": 0.000004, |
|
"litellm_provider": "openai", |
|
"mode": "chat" |
|
}, |
|
"text-davinci-003": { |
|
"max_tokens": 4097, |
|
"input_cost_per_token": 0.000002, |
|
"output_cost_per_token": 0.000002, |
|
"litellm_provider": "text-completion-openai", |
|
"mode": "completion" |
|
}, |
|
"text-curie-001": { |
|
"max_tokens": 2049, |
|
"input_cost_per_token": 0.000002, |
|
"output_cost_per_token": 0.000002, |
|
"litellm_provider": "text-completion-openai", |
|
"mode": "completion" |
|
}, |
|
"text-babbage-001": { |
|
"max_tokens": 2049, |
|
"input_cost_per_token": 0.0000004, |
|
"output_cost_per_token": 0.0000004, |
|
"litellm_provider": "text-completion-openai", |
|
"mode": "completion" |
|
}, |
|
"text-ada-001": { |
|
"max_tokens": 2049, |
|
"input_cost_per_token": 0.0000004, |
|
"output_cost_per_token": 0.0000004, |
|
"litellm_provider": "text-completion-openai", |
|
"mode": "completion" |
|
}, |
|
"babbage-002": { |
|
"max_tokens": 16384, |
|
"input_cost_per_token": 0.0000004, |
|
"output_cost_per_token": 0.0000004, |
|
"litellm_provider": "text-completion-openai", |
|
"mode": "completion" |
|
}, |
|
"davinci-002": { |
|
"max_tokens": 16384, |
|
"input_cost_per_token": 0.000002, |
|
"output_cost_per_token": 0.000002, |
|
"litellm_provider": "text-completion-openai", |
|
"mode": "completion" |
|
}, |
|
"gpt-3.5-turbo-instruct": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.0000015, |
|
"output_cost_per_token": 0.000002, |
|
"litellm_provider": "text-completion-openai", |
|
"mode": "completion" |
|
}, |
|
"claude-instant-1": { |
|
"max_tokens": 100000, |
|
"input_cost_per_token": 0.00000163, |
|
"output_cost_per_token": 0.00000551, |
|
"litellm_provider": "anthropic", |
|
"mode": "chat" |
|
}, |
|
"claude-instant-1.2": { |
|
"max_tokens": 100000, |
|
"input_cost_per_token": 0.00000163, |
|
"output_cost_per_token": 0.00000551, |
|
"litellm_provider": "anthropic", |
|
"mode": "chat" |
|
}, |
|
"claude-2": { |
|
"max_tokens": 100000, |
|
"input_cost_per_token": 0.00001102, |
|
"output_cost_per_token": 0.00003268, |
|
"litellm_provider": "anthropic", |
|
"mode": "chat" |
|
}, |
|
"text-bison": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-text-models", |
|
"mode": "completion" |
|
}, |
|
"text-bison@001": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-text-models", |
|
"mode": "completion" |
|
}, |
|
"chat-bison": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-chat-models", |
|
"mode": "chat" |
|
}, |
|
"chat-bison@001": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-chat-models", |
|
"mode": "chat" |
|
}, |
|
"chat-bison-32k": { |
|
"max_tokens": 32000, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-chat-models", |
|
"mode": "chat" |
|
}, |
|
"code-bison": { |
|
"max_tokens": 6144, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-code-text-models", |
|
"mode": "chat" |
|
}, |
|
"code-bison@001": { |
|
"max_tokens": 6144, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-code-text-models", |
|
"mode": "completion" |
|
}, |
|
"code-gecko@001": { |
|
"max_tokens": 2048, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-chat-models", |
|
"mode": "completion" |
|
}, |
|
"code-gecko@latest": { |
|
"max_tokens": 2048, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-chat-models", |
|
"mode": "completion" |
|
}, |
|
"codechat-bison": { |
|
"max_tokens": 6144, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-code-chat-models", |
|
"mode": "chat" |
|
}, |
|
"codechat-bison@001": { |
|
"max_tokens": 6144, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-code-chat-models", |
|
"mode": "chat" |
|
}, |
|
"codechat-bison-32k": { |
|
"max_tokens": 32000, |
|
"input_cost_per_token": 0.000000125, |
|
"output_cost_per_token": 0.000000125, |
|
"litellm_provider": "vertex_ai-chat-models", |
|
"mode": "chat" |
|
}, |
|
"command-nightly": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.000015, |
|
"output_cost_per_token": 0.000015, |
|
"litellm_provider": "cohere", |
|
"mode": "completion" |
|
}, |
|
"command": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.000015, |
|
"output_cost_per_token": 0.000015, |
|
"litellm_provider": "cohere", |
|
"mode": "completion" |
|
}, |
|
"command-light": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.000015, |
|
"output_cost_per_token": 0.000015, |
|
"litellm_provider": "cohere", |
|
"mode": "completion" |
|
}, |
|
"command-medium-beta": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.000015, |
|
"output_cost_per_token": 0.000015, |
|
"litellm_provider": "cohere", |
|
"mode": "completion" |
|
}, |
|
"command-xlarge-beta": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.000015, |
|
"output_cost_per_token": 0.000015, |
|
"litellm_provider": "cohere", |
|
"mode": "completion" |
|
}, |
|
"replicate/llama-2-70b-chat:2c1608e18606fad2812020dc541930f2d0495ce32eee50074220b87300bc16e1": { |
|
"max_tokens": 4096, |
|
"litellm_provider": "replicate", |
|
"mode": "chat" |
|
}, |
|
"openrouter/openai/gpt-3.5-turbo": { |
|
"max_tokens": 4095, |
|
"input_cost_per_token": 0.0000015, |
|
"output_cost_per_token": 0.000002, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/openai/gpt-3.5-turbo-16k": { |
|
"max_tokens": 16383, |
|
"input_cost_per_token": 0.000003, |
|
"output_cost_per_token": 0.000004, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/openai/gpt-4": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.00003, |
|
"output_cost_per_token": 0.00006, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/anthropic/claude-instant-v1": { |
|
"max_tokens": 100000, |
|
"input_cost_per_token": 0.00000163, |
|
"output_cost_per_token": 0.00000551, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/anthropic/claude-2": { |
|
"max_tokens": 100000, |
|
"input_cost_per_token": 0.00001102, |
|
"output_cost_per_token": 0.00003268, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/google/palm-2-chat-bison": { |
|
"max_tokens": 8000, |
|
"input_cost_per_token": 0.0000005, |
|
"output_cost_per_token": 0.0000005, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/google/palm-2-codechat-bison": { |
|
"max_tokens": 8000, |
|
"input_cost_per_token": 0.0000005, |
|
"output_cost_per_token": 0.0000005, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/meta-llama/llama-2-13b-chat": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0000002, |
|
"output_cost_per_token": 0.0000002, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/meta-llama/llama-2-70b-chat": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0000015, |
|
"output_cost_per_token": 0.0000015, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/meta-llama/codellama-34b-instruct": { |
|
"max_tokens": 8096, |
|
"input_cost_per_token": 0.0000005, |
|
"output_cost_per_token": 0.0000005, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/nousresearch/nous-hermes-llama2-13b": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0000002, |
|
"output_cost_per_token": 0.0000002, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/mancer/weaver": { |
|
"max_tokens": 8000, |
|
"input_cost_per_token": 0.000005625, |
|
"output_cost_per_token": 0.000005625, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/gryphe/mythomax-l2-13b": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.000001875, |
|
"output_cost_per_token": 0.000001875, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/jondurbin/airoboros-l2-70b-2.1": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.000013875, |
|
"output_cost_per_token": 0.000013875, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/undi95/remm-slerp-l2-13b": { |
|
"max_tokens": 6144, |
|
"input_cost_per_token": 0.000001875, |
|
"output_cost_per_token": 0.000001875, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/pygmalionai/mythalion-13b": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.000001875, |
|
"output_cost_per_token": 0.000001875, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"openrouter/mistralai/mistral-7b-instruct": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0, |
|
"output_cost_per_token": 0.0, |
|
"litellm_provider": "openrouter", |
|
"mode": "chat" |
|
}, |
|
"j2-ultra": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.000015, |
|
"output_cost_per_token": 0.000015, |
|
"litellm_provider": "ai21", |
|
"mode": "completion" |
|
}, |
|
"j2-mid": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.00001, |
|
"output_cost_per_token": 0.00001, |
|
"litellm_provider": "ai21", |
|
"mode": "completion" |
|
}, |
|
"j2-light": { |
|
"max_tokens": 8192, |
|
"input_cost_per_token": 0.000003, |
|
"output_cost_per_token": 0.000003, |
|
"litellm_provider": "ai21", |
|
"mode": "completion" |
|
}, |
|
"dolphin": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.00002, |
|
"output_cost_per_token": 0.00002, |
|
"litellm_provider": "nlp_cloud", |
|
"mode": "completion" |
|
}, |
|
"chatdolphin": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.00002, |
|
"output_cost_per_token": 0.00002, |
|
"litellm_provider": "nlp_cloud", |
|
"mode": "chat" |
|
}, |
|
"luminous-base": { |
|
"max_tokens": 2048, |
|
"input_cost_per_token": 0.00003, |
|
"output_cost_per_token": 0.000033, |
|
"litellm_provider": "aleph_alpha", |
|
"mode": "completion" |
|
}, |
|
"luminous-base-control": { |
|
"max_tokens": 2048, |
|
"input_cost_per_token": 0.0000375, |
|
"output_cost_per_token": 0.00004125, |
|
"litellm_provider": "aleph_alpha", |
|
"mode": "chat" |
|
}, |
|
"luminous-extended": { |
|
"max_tokens": 2048, |
|
"input_cost_per_token": 0.000045, |
|
"output_cost_per_token": 0.0000495, |
|
"litellm_provider": "aleph_alpha", |
|
"mode": "completion" |
|
}, |
|
"luminous-extended-control": { |
|
"max_tokens": 2048, |
|
"input_cost_per_token": 0.00005625, |
|
"output_cost_per_token": 0.000061875, |
|
"litellm_provider": "aleph_alpha", |
|
"mode": "chat" |
|
}, |
|
"luminous-supreme": { |
|
"max_tokens": 2048, |
|
"input_cost_per_token": 0.000175, |
|
"output_cost_per_token": 0.0001925, |
|
"litellm_provider": "aleph_alpha", |
|
"mode": "completion" |
|
}, |
|
"luminous-supreme-control": { |
|
"max_tokens": 2048, |
|
"input_cost_per_token": 0.00021875, |
|
"output_cost_per_token": 0.000240625, |
|
"litellm_provider": "aleph_alpha", |
|
"mode": "chat" |
|
}, |
|
"ai21.j2-mid-v1": { |
|
"max_tokens": 8191, |
|
"input_cost_per_token": 0.0000125, |
|
"output_cost_per_token": 0.0000125, |
|
"litellm_provider": "bedrock", |
|
"mode": "chat" |
|
}, |
|
"ai21.j2-ultra-v1": { |
|
"max_tokens": 8191, |
|
"input_cost_per_token": 0.0000188, |
|
"output_cost_per_token": 0.0000188, |
|
"litellm_provider": "bedrock", |
|
"mode": "chat" |
|
}, |
|
"amazon.titan-text-lite-v1": { |
|
"max_tokens": 8000, |
|
"input_cost_per_token": 0.0000003, |
|
"output_cost_per_token": 0.0000004, |
|
"litellm_provider": "bedrock", |
|
"mode": "chat" |
|
}, |
|
"amazon.titan-text-express-v1": { |
|
"max_tokens": 8000, |
|
"input_cost_per_token": 0.0000013, |
|
"output_cost_per_token": 0.0000017, |
|
"litellm_provider": "bedrock", |
|
"mode": "chat" |
|
}, |
|
"anthropic.claude-v1": { |
|
"max_tokens": 100000, |
|
"input_cost_per_token": 0.00001102, |
|
"output_cost_per_token": 0.00003268, |
|
"litellm_provider": "bedrock", |
|
"mode": "chat" |
|
}, |
|
"anthropic.claude-v2": { |
|
"max_tokens": 100000, |
|
"input_cost_per_token": 0.00001102, |
|
"output_cost_per_token": 0.00003268, |
|
"litellm_provider": "bedrock", |
|
"mode": "chat" |
|
}, |
|
"anthropic.claude-instant-v1": { |
|
"max_tokens": 100000, |
|
"input_cost_per_token": 0.00000163, |
|
"output_cost_per_token": 0.00000551, |
|
"litellm_provider": "bedrock", |
|
"mode": "chat" |
|
}, |
|
"cohere.command-text-v14": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0000015, |
|
"output_cost_per_token": 0.0000020, |
|
"litellm_provider": "bedrock", |
|
"mode": "chat" |
|
}, |
|
"meta.llama2-13b-chat-v1": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.00000075, |
|
"output_cost_per_token": 0.000001, |
|
"litellm_provider": "bedrock", |
|
"mode": "chat" |
|
}, |
|
"together-ai-up-to-3b": { |
|
"input_cost_per_token": 0.0000001, |
|
"output_cost_per_token": 0.0000001 |
|
}, |
|
"together-ai-3.1b-7b": { |
|
"input_cost_per_token": 0.0000002, |
|
"output_cost_per_token": 0.0000002 |
|
}, |
|
"together-ai-7.1b-20b": { |
|
"max_tokens": 1000, |
|
"input_cost_per_token": 0.0000004, |
|
"output_cost_per_token": 0.0000004 |
|
}, |
|
"together-ai-20.1b-40b": { |
|
"input_cost_per_token": 0.000001, |
|
"output_cost_per_token": 0.000001 |
|
}, |
|
"together-ai-40.1b-70b": { |
|
"input_cost_per_token": 0.000003, |
|
"output_cost_per_token": 0.000003 |
|
}, |
|
"ollama/llama2": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0, |
|
"output_cost_per_token": 0.0, |
|
"litellm_provider": "ollama", |
|
"mode": "completion" |
|
}, |
|
"ollama/llama2:13b": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0, |
|
"output_cost_per_token": 0.0, |
|
"litellm_provider": "ollama", |
|
"mode": "completion" |
|
}, |
|
"ollama/llama2:70b": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0, |
|
"output_cost_per_token": 0.0, |
|
"litellm_provider": "ollama", |
|
"mode": "completion" |
|
}, |
|
"ollama/llama2-uncensored": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0, |
|
"output_cost_per_token": 0.0, |
|
"litellm_provider": "ollama", |
|
"mode": "completion" |
|
}, |
|
"deepinfra/meta-llama/Llama-2-70b-chat-hf": { |
|
"max_tokens": 6144, |
|
"input_cost_per_token": 0.000001875, |
|
"output_cost_per_token": 0.000001875, |
|
"litellm_provider": "deepinfra", |
|
"mode": "chat" |
|
}, |
|
"deepinfra/codellama/CodeLlama-34b-Instruct-hf": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0000006, |
|
"output_cost_per_token": 0.0000006, |
|
"litellm_provider": "deepinfra", |
|
"mode": "chat" |
|
}, |
|
"deepinfra/meta-llama/Llama-2-13b-chat-hf": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.00000035, |
|
"output_cost_per_token": 0.00000035, |
|
"litellm_provider": "deepinfra", |
|
"mode": "chat" |
|
}, |
|
"deepinfra/meta-llama/Llama-2-7b-chat-hf": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0000002, |
|
"output_cost_per_token": 0.0000002, |
|
"litellm_provider": "deepinfra", |
|
"mode": "chat" |
|
}, |
|
"deepinfra/mistralai/Mistral-7B-Instruct-v0.1": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0000002, |
|
"output_cost_per_token": 0.0000002, |
|
"litellm_provider": "deepinfra", |
|
"mode": "chat" |
|
}, |
|
"deepinfra/jondurbin/airoboros-l2-70b-gpt4-1.4.1": { |
|
"max_tokens": 4096, |
|
"input_cost_per_token": 0.0000007, |
|
"output_cost_per_token": 0.00000095, |
|
"litellm_provider": "deepinfra", |
|
"mode": "chat" |
|
} |
|
} |
|
|