ragflow / conf /llm_factories.json
zxsipola123456's picture
Upload 769 files
ab2ded1 verified
raw
history blame
88.2 kB
{
"factory_llm_infos": [
{
"name": "OpenAI",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION",
"status": "1",
"llm": [
{
"llm_name": "gpt-4o-mini",
"tags": "LLM,CHAT,128K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "gpt-4o",
"tags": "LLM,CHAT,128K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "gpt-3.5-turbo",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "gpt-3.5-turbo-16k-0613",
"tags": "LLM,CHAT,16k",
"max_tokens": 16385,
"model_type": "chat"
},
{
"llm_name": "text-embedding-ada-002",
"tags": "TEXT EMBEDDING,8K",
"max_tokens": 8191,
"model_type": "embedding"
},
{
"llm_name": "text-embedding-3-small",
"tags": "TEXT EMBEDDING,8K",
"max_tokens": 8191,
"model_type": "embedding"
},
{
"llm_name": "text-embedding-3-large",
"tags": "TEXT EMBEDDING,8K",
"max_tokens": 8191,
"model_type": "embedding"
},
{
"llm_name": "whisper-1",
"tags": "SPEECH2TEXT",
"max_tokens": 26214400,
"model_type": "speech2text"
},
{
"llm_name": "gpt-4",
"tags": "LLM,CHAT,8K",
"max_tokens": 8191,
"model_type": "chat"
},
{
"llm_name": "gpt-4-turbo",
"tags": "LLM,CHAT,8K",
"max_tokens": 8191,
"model_type": "chat"
},
{
"llm_name": "gpt-4-32k",
"tags": "LLM,CHAT,32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "gpt-4-vision-preview",
"tags": "LLM,CHAT,IMAGE2TEXT",
"max_tokens": 765,
"model_type": "image2text"
}
]
},
{
"name": "Tongyi-Qianwen",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION",
"status": "1",
"llm": [
{
"llm_name": "qwen-turbo",
"tags": "LLM,CHAT,8K",
"max_tokens": 8191,
"model_type": "chat"
},
{
"llm_name": "qwen-plus",
"tags": "LLM,CHAT,32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "text-embedding-v2",
"tags": "TEXT EMBEDDING,2K",
"max_tokens": 2048,
"model_type": "embedding"
},
{
"llm_name": "paraformer-realtime-8k-v1",
"tags": "SPEECH2TEXT",
"max_tokens": 26214400,
"model_type": "speech2text"
},
{
"llm_name": "qwen-vl-max",
"tags": "LLM,CHAT,IMAGE2TEXT",
"max_tokens": 765,
"model_type": "image2text"
},
{
"llm_name": "qwen-vl-plus",
"tags": "LLM,CHAT,IMAGE2TEXT",
"max_tokens": 765,
"model_type": "image2text"
}
]
},
{
"name": "ZHIPU-AI",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION",
"status": "1",
"llm": [
{
"llm_name": "glm-3-turbo",
"tags": "LLM,CHAT,",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "glm-4",
"tags": "LLM,CHAT,",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "glm-4v",
"tags": "LLM,CHAT,IMAGE2TEXT",
"max_tokens": 2000,
"model_type": "image2text"
},
{
"llm_name": "embedding-2",
"tags": "TEXT EMBEDDING",
"max_tokens": 512,
"model_type": "embedding"
}
]
},
{
"name": "Ollama",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION",
"status": "1",
"llm": []
},
{
"name": "LocalAI",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION",
"status": "1",
"llm": []
},
{
"name": "OpenAI-API-Compatible",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION",
"status": "1",
"llm": []
},
{
"name": "Moonshot",
"logo": "",
"tags": "LLM,TEXT EMBEDDING",
"status": "1",
"llm": [
{
"llm_name": "moonshot-v1-8k",
"tags": "LLM,CHAT,",
"max_tokens": 7900,
"model_type": "chat"
},
{
"llm_name": "moonshot-v1-32k",
"tags": "LLM,CHAT,",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "moonshot-v1-128k",
"tags": "LLM,CHAT",
"max_tokens": 128000,
"model_type": "chat"
}
]
},
{
"name": "FastEmbed",
"logo": "",
"tags": "TEXT EMBEDDING",
"status": "1",
"llm": [
{
"llm_name": "BAAI/bge-small-en-v1.5",
"tags": "TEXT EMBEDDING,",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "BAAI/bge-small-zh-v1.5",
"tags": "TEXT EMBEDDING,",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "BAAI/bge-base-en-v1.5",
"tags": "TEXT EMBEDDING,",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "BAAI/bge-large-en-v1.5",
"tags": "TEXT EMBEDDING,",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "sentence-transformers/all-MiniLM-L6-v2",
"tags": "TEXT EMBEDDING,",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "nomic-ai/nomic-embed-text-v1.5",
"tags": "TEXT EMBEDDING,",
"max_tokens": 8192,
"model_type": "embedding"
},
{
"llm_name": "jinaai/jina-embeddings-v2-small-en",
"tags": "TEXT EMBEDDING,",
"max_tokens": 2147483648,
"model_type": "embedding"
},
{
"llm_name": "jinaai/jina-embeddings-v2-base-en",
"tags": "TEXT EMBEDDING,",
"max_tokens": 2147483648,
"model_type": "embedding"
}
]
},
{
"name": "Xinference",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION,TEXT RE-RANK",
"status": "1",
"llm": []
},
{
"name": "Youdao",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION",
"status": "1",
"llm": [
{
"llm_name": "maidalun1020/bce-embedding-base_v1",
"tags": "TEXT EMBEDDING,",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "maidalun1020/bce-reranker-base_v1",
"tags": "RE-RANK, 512",
"max_tokens": 512,
"model_type": "rerank"
}
]
},
{
"name": "DeepSeek",
"logo": "",
"tags": "LLM",
"status": "1",
"llm": [
{
"llm_name": "deepseek-chat",
"tags": "LLM,CHAT,",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "deepseek-coder",
"tags": "LLM,CHAT,",
"max_tokens": 16385,
"model_type": "chat"
}
]
},
{
"name": "VolcEngine",
"logo": "",
"tags": "LLM, TEXT EMBEDDING",
"status": "1",
"llm": [
{
"llm_name": "Skylark2-pro-32k",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "Skylark2-pro-4k",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
}
]
},
{
"name": "BaiChuan",
"logo": "",
"tags": "LLM,TEXT EMBEDDING",
"status": "1",
"llm": [
{
"llm_name": "Baichuan2-Turbo",
"tags": "LLM,CHAT,32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "Baichuan2-Turbo-192k",
"tags": "LLM,CHAT,192K",
"max_tokens": 196608,
"model_type": "chat"
},
{
"llm_name": "Baichuan3-Turbo",
"tags": "LLM,CHAT,32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "Baichuan3-Turbo-128k",
"tags": "LLM,CHAT,128K",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "Baichuan4",
"tags": "LLM,CHAT,128K",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "Baichuan-Text-Embedding",
"tags": "TEXT EMBEDDING",
"max_tokens": 512,
"model_type": "embedding"
}
]
},
{
"name": "Jina",
"logo": "",
"tags": "TEXT EMBEDDING, TEXT RE-RANK",
"status": "1",
"llm": [
{
"llm_name": "jina-reranker-v1-base-en",
"tags": "RE-RANK,8k",
"max_tokens": 8196,
"model_type": "rerank"
},
{
"llm_name": "jina-reranker-v1-turbo-en",
"tags": "RE-RANK,8k",
"max_tokens": 8196,
"model_type": "rerank"
},
{
"llm_name": "jina-reranker-v1-tiny-en",
"tags": "RE-RANK,8k",
"max_tokens": 8196,
"model_type": "rerank"
},
{
"llm_name": "jina-colbert-v1-en",
"tags": "RE-RANK,8k",
"max_tokens": 8196,
"model_type": "rerank"
},
{
"llm_name": "jina-embeddings-v2-base-en",
"tags": "TEXT EMBEDDING",
"max_tokens": 8196,
"model_type": "embedding"
},
{
"llm_name": "jina-embeddings-v2-base-de",
"tags": "TEXT EMBEDDING",
"max_tokens": 8196,
"model_type": "embedding"
},
{
"llm_name": "jina-embeddings-v2-base-es",
"tags": "TEXT EMBEDDING",
"max_tokens": 8196,
"model_type": "embedding"
},
{
"llm_name": "jina-embeddings-v2-base-code",
"tags": "TEXT EMBEDDING",
"max_tokens": 8196,
"model_type": "embedding"
},
{
"llm_name": "jina-embeddings-v2-base-zh",
"tags": "TEXT EMBEDDING",
"max_tokens": 8196,
"model_type": "embedding"
}
]
},
{
"name": "BAAI",
"logo": "",
"tags": "TEXT EMBEDDING, TEXT RE-RANK",
"status": "1",
"llm": [
{
"llm_name": "BAAI/bge-large-zh-v1.5",
"tags": "TEXT EMBEDDING,",
"max_tokens": 1024,
"model_type": "embedding"
},
{
"llm_name": "BAAI/bge-reranker-v2-m3",
"tags": "RE-RANK,2k",
"max_tokens": 2048,
"model_type": "rerank"
}
]
},
{
"name": "MiniMax",
"logo": "",
"tags": "LLM,TEXT EMBEDDING",
"status": "1",
"llm": [
{
"llm_name": "abab6.5-chat",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "abab6.5s-chat",
"tags": "LLM,CHAT,245k",
"max_tokens": 245760,
"model_type": "chat"
},
{
"llm_name": "abab6.5t-chat",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "abab6.5g-chat",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "abab5.5s-chat",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
}
]
},
{
"name": "Mistral",
"logo": "",
"tags": "LLM,TEXT EMBEDDING",
"status": "1",
"llm": [
{
"llm_name": "open-mixtral-8x22b",
"tags": "LLM,CHAT,64k",
"max_tokens": 64000,
"model_type": "chat"
},
{
"llm_name": "open-mixtral-8x7b",
"tags": "LLM,CHAT,32k",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "open-mistral-7b",
"tags": "LLM,CHAT,32k",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "mistral-large-latest",
"tags": "LLM,CHAT,32k",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "mistral-small-latest",
"tags": "LLM,CHAT,32k",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "mistral-medium-latest",
"tags": "LLM,CHAT,32k",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "codestral-latest",
"tags": "LLM,CHAT,32k",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "mistral-embed",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "embedding"
}
]
},
{
"name": "Azure-OpenAI",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION",
"status": "1",
"llm": [
{
"llm_name": "azure-gpt-4o-mini",
"tags": "LLM,CHAT,128K",
"max_tokens": 128000,
"model_type": "image2text"
},
{
"llm_name": "azure-gpt-4o",
"tags": "LLM,CHAT,128K",
"max_tokens": 128000,
"model_type": "chat,image2text"
},
{
"llm_name": "azure-gpt-35-turbo",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "azure-gpt-35-turbo-16k",
"tags": "LLM,CHAT,16k",
"max_tokens": 16385,
"model_type": "chat"
},
{
"llm_name": "azure-text-embedding-ada-002",
"tags": "TEXT EMBEDDING,8K",
"max_tokens": 8191,
"model_type": "embedding"
},
{
"llm_name": "azure-text-embedding-3-small",
"tags": "TEXT EMBEDDING,8K",
"max_tokens": 8191,
"model_type": "embedding"
},
{
"llm_name": "azure-text-embedding-3-large",
"tags": "TEXT EMBEDDING,8K",
"max_tokens": 8191,
"model_type": "embedding"
},
{
"llm_name": "azure-whisper-1",
"tags": "SPEECH2TEXT",
"max_tokens": 26214400,
"model_type": "speech2text"
},
{
"llm_name": "azure-gpt-4",
"tags": "LLM,CHAT,8K",
"max_tokens": 8191,
"model_type": "chat"
},
{
"llm_name": "azure-gpt-4-turbo",
"tags": "LLM,CHAT,8K",
"max_tokens": 8191,
"model_type": "chat"
},
{
"llm_name": "azure-gpt-4-32k",
"tags": "LLM,CHAT,32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "azure-gpt-4-vision-preview",
"tags": "LLM,CHAT,IMAGE2TEXT",
"max_tokens": 765,
"model_type": "image2text"
}
]
},
{
"name": "Bedrock",
"logo": "",
"tags": "LLM,TEXT EMBEDDING",
"status": "1",
"llm": [
{
"llm_name": "ai21.j2-ultra-v1",
"tags": "LLM,CHAT,8k",
"max_tokens": 8191,
"model_type": "chat"
},
{
"llm_name": "ai21.j2-mid-v1",
"tags": "LLM,CHAT,8k",
"max_tokens": 8191,
"model_type": "chat"
},
{
"llm_name": "cohere.command-text-v14",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "cohere.command-light-text-v14",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "cohere.command-r-v1:0",
"tags": "LLM,CHAT,128k",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "cohere.command-r-plus-v1:0",
"tags": "LLM,CHAT,128k",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "anthropic.claude-v2",
"tags": "LLM,CHAT,100k",
"max_tokens": 102400,
"model_type": "chat"
},
{
"llm_name": "anthropic.claude-v2:1",
"tags": "LLM,CHAT,200k",
"max_tokens": 204800,
"model_type": "chat"
},
{
"llm_name": "anthropic.claude-3-sonnet-20240229-v1:0",
"tags": "LLM,CHAT,200k",
"max_tokens": 204800,
"model_type": "chat"
},
{
"llm_name": "anthropic.claude-3-5-sonnet-20240620-v1:0",
"tags": "LLM,CHAT,200k",
"max_tokens": 204800,
"model_type": "chat"
},
{
"llm_name": "anthropic.claude-3-haiku-20240307-v1:0",
"tags": "LLM,CHAT,200k",
"max_tokens": 204800,
"model_type": "chat"
},
{
"llm_name": "anthropic.claude-3-opus-20240229-v1:0",
"tags": "LLM,CHAT,200k",
"max_tokens": 204800,
"model_type": "chat"
},
{
"llm_name": "anthropic.claude-instant-v1",
"tags": "LLM,CHAT,100k",
"max_tokens": 102400,
"model_type": "chat"
},
{
"llm_name": "amazon.titan-text-express-v1",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "amazon.titan-text-premier-v1:0",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "amazon.titan-text-lite-v1",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "meta.llama2-13b-chat-v1",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "meta.llama2-70b-chat-v1",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "meta.llama3-8b-instruct-v1:0",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "meta.llama3-70b-instruct-v1:0",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "mistral.mistral-7b-instruct-v0:2",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "mistral.mixtral-8x7b-instruct-v0:1",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "mistral.mistral-large-2402-v1:0",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "mistral.mistral-small-2402-v1:0",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "amazon.titan-embed-text-v2:0",
"tags": "TEXT EMBEDDING",
"max_tokens": 8192,
"model_type": "embedding"
},
{
"llm_name": "cohere.embed-english-v3",
"tags": "TEXT EMBEDDING",
"max_tokens": 2048,
"model_type": "embedding"
},
{
"llm_name": "cohere.embed-multilingual-v3",
"tags": "TEXT EMBEDDING",
"max_tokens": 2048,
"model_type": "embedding"
}
]
},
{
"name": "Gemini",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,IMAGE2TEXT",
"status": "1",
"llm": [
{
"llm_name": "gemini-1.5-pro-latest",
"tags": "LLM,CHAT,1024K",
"max_tokens": 1048576,
"model_type": "chat"
},
{
"llm_name": "gemini-1.5-flash-latest",
"tags": "LLM,CHAT,1024K",
"max_tokens": 1048576,
"model_type": "chat"
},
{
"llm_name": "gemini-1.0-pro",
"tags": "LLM,CHAT,30K",
"max_tokens": 30720,
"model_type": "chat"
},
{
"llm_name": "gemini-1.0-pro-vision-latest",
"tags": "LLM,IMAGE2TEXT,12K",
"max_tokens": 12288,
"model_type": "image2text"
},
{
"llm_name": "text-embedding-004",
"tags": "TEXT EMBEDDING",
"max_tokens": 2048,
"model_type": "embedding"
}
]
},
{
"name": "Groq",
"logo": "",
"tags": "LLM",
"status": "1",
"llm": [
{
"llm_name": "gemma-7b-it",
"tags": "LLM,CHAT,15k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "gemma2-9b-it",
"tags": "LLM,CHAT,15k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "llama3-70b-8192",
"tags": "LLM,CHAT,6k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "llama3-8b-8192",
"tags": "LLM,CHAT,30k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "mixtral-8x7b-32768",
"tags": "LLM,CHAT,5k",
"max_tokens": 32768,
"model_type": "chat"
}
]
},
{
"name": "OpenRouter",
"logo": "",
"tags": "LLM,IMAGE2TEXT",
"status": "1",
"llm": [
{
"llm_name": "nousresearch/hermes-2-theta-llama-3-8b",
"tags": "LLM CHAT 16K",
"max_tokens": 16384,
"model_type": "chat"
},
{
"llm_name": "alpindale/magnum-72b",
"tags": "LLM CHAT 16K",
"max_tokens": 16384,
"model_type": "chat"
},
{
"llm_name": "google/gemma-2-9b-it",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "google/gemma-2-9b-it:free",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "sao10k/l3-stheno-8b",
"tags": "LLM CHAT 32K",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "openrouter/flavor-of-the-week",
"tags": "LLM CHAT 32K",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "ai21/jamba-instruct",
"tags": "LLM CHAT 250K",
"max_tokens": 256000,
"model_type": "chat"
},
{
"llm_name": "nvidia/nemotron-4-340b-instruct",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-3.5-sonnet",
"tags": "LLM IMAGE2TEXT 195K",
"max_tokens": 200000,
"model_type": "image2text"
},
{
"llm_name": "anthropic/claude-3.5-sonnet:beta",
"tags": "LLM IMAGE2TEXT 195K",
"max_tokens": 200000,
"model_type": "image2text"
},
{
"llm_name": "sao10k/l3-euryale-70b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "microsoft/phi-3-medium-4k-instruct",
"tags": "LLM CHAT 4K",
"max_tokens": 4000,
"model_type": "chat"
},
{
"llm_name": "cognitivecomputations/dolphin-mixtral-8x22b",
"tags": "LLM CHAT 64K",
"max_tokens": 65536,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen-2-72b-instruct",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "openchat/openchat-8b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-7b-instruct",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-7b-instruct-v0.3",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "nousresearch/hermes-2-pro-llama-3-8b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "microsoft/phi-3-mini-128k-instruct",
"tags": "LLM CHAT 125K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "microsoft/phi-3-mini-128k-instruct:free",
"tags": "LLM CHAT 125K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "microsoft/phi-3-medium-128k-instruct",
"tags": "LLM CHAT 125K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "microsoft/phi-3-medium-128k-instruct:free",
"tags": "LLM CHAT 125K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "neversleep/llama-3-lumimaid-70b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "google/gemini-flash-1.5",
"tags": "LLM IMAGE2TEXT 2734K",
"max_tokens": 2800000,
"model_type": "image2text"
},
{
"llm_name": "perplexity/llama-3-sonar-small-32k-chat",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "perplexity/llama-3-sonar-small-32k-online",
"tags": "LLM CHAT 28K",
"max_tokens": 28000,
"model_type": "chat"
},
{
"llm_name": "perplexity/llama-3-sonar-large-32k-chat",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "perplexity/llama-3-sonar-large-32k-online",
"tags": "LLM CHAT 28K",
"max_tokens": 28000,
"model_type": "chat"
},
{
"llm_name": "deepseek/deepseek-chat",
"tags": "LLM CHAT 125K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "deepseek/deepseek-coder",
"tags": "LLM CHAT 125K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-4o",
"tags": "LLM IMAGE2TEXT 125K",
"max_tokens": 128000,
"model_type": "image2text"
},
{
"llm_name": "openai/gpt-4o-2024-05-13",
"tags": "LLM IMAGE2TEXT 125K",
"max_tokens": 128000,
"model_type": "image2text"
},
{
"llm_name": "meta-llama/llama-3-8b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-3-70b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-guard-2-8b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "liuhaotian/llava-yi-34b",
"tags": "LLM IMAGE2TEXT 4K",
"max_tokens": 4096,
"model_type": "image2text"
},
{
"llm_name": "allenai/olmo-7b-instruct",
"tags": "LLM CHAT 2K",
"max_tokens": 2048,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen-110b-chat",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen-72b-chat",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen-32b-chat",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen-14b-chat",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen-7b-chat",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "qwen/qwen-4b-chat",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-3-8b-instruct:free",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "neversleep/llama-3-lumimaid-8b",
"tags": "LLM CHAT 24K",
"max_tokens": 24576,
"model_type": "chat"
},
{
"llm_name": "neversleep/llama-3-lumimaid-8b:extended",
"tags": "LLM CHAT 24K",
"max_tokens": 24576,
"model_type": "chat"
},
{
"llm_name": "snowflake/snowflake-arctic-instruct",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "fireworks/firellava-13b",
"tags": "LLM IMAGE2TEXT 4K",
"max_tokens": 4096,
"model_type": "image2text"
},
{
"llm_name": "lynn/soliloquy-l3",
"tags": "LLM CHAT 24K",
"max_tokens": 24576,
"model_type": "chat"
},
{
"llm_name": "sao10k/fimbulvetr-11b-v2",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-3-8b-instruct:extended",
"tags": "LLM CHAT 16K",
"max_tokens": 16384,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-3-8b-instruct:nitro",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-3-70b-instruct:nitro",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-3-8b-instruct",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-3-70b-instruct",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "mistralai/mixtral-8x22b-instruct",
"tags": "LLM CHAT 64K",
"max_tokens": 65536,
"model_type": "chat"
},
{
"llm_name": "microsoft/wizardlm-2-8x22b",
"tags": "LLM CHAT 64K",
"max_tokens": 65536,
"model_type": "chat"
},
{
"llm_name": "microsoft/wizardlm-2-7b",
"tags": "LLM CHAT 32K",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "undi95/toppy-m-7b:nitro",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "mistralai/mixtral-8x22b",
"tags": "LLM CHAT 64K",
"max_tokens": 65536,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-4-turbo",
"tags": "LLM IMAGE2TEXT 125K",
"max_tokens": 128000,
"model_type": "image2text"
},
{
"llm_name": "google/gemini-pro-1.5",
"tags": "LLM IMAGE2TEXT 2734K",
"max_tokens": 2800000,
"model_type": "image2text"
},
{
"llm_name": "cohere/command-r-plus",
"tags": "LLM CHAT 125K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "databricks/dbrx-instruct",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "sophosympatheia/midnight-rose-70b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "cohere/command",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "cohere/command-r",
"tags": "LLM CHAT 125K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-3-haiku",
"tags": "LLM IMAGE2TEXT 195K",
"max_tokens": 200000,
"model_type": "image2text"
},
{
"llm_name": "anthropic/claude-3-haiku:beta",
"tags": "LLM IMAGE2TEXT 195K",
"max_tokens": 200000,
"model_type": "image2text"
},
{
"llm_name": "google/gemma-7b-it:nitro",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "mistralai/mixtral-8x7b-instruct:nitro",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-7b-instruct:nitro",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-2-70b-chat:nitro",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "gryphe/mythomax-l2-13b:nitro",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-3-opus",
"tags": "LLM IMAGE2TEXT 195K",
"max_tokens": 200000,
"model_type": "image2text"
},
{
"llm_name": "anthropic/claude-3-sonnet",
"tags": "LLM IMAGE2TEXT 195K",
"max_tokens": 200000,
"model_type": "image2text"
},
{
"llm_name": "anthropic/claude-3-opus:beta",
"tags": "LLM IMAGE2TEXT 195K",
"max_tokens": 200000,
"model_type": "image2text"
},
{
"llm_name": "anthropic/claude-3-sonnet:beta",
"tags": "LLM IMAGE2TEXT 195K",
"max_tokens": 200000,
"model_type": "image2text"
},
{
"llm_name": "mistralai/mistral-large",
"tags": "LLM CHAT 32K",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "google/gemma-7b-it",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "google/gemma-7b-it:free",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "nousresearch/nous-hermes-2-mistral-7b-dpo",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "meta-llama/codellama-70b-instruct",
"tags": "LLM CHAT 2K",
"max_tokens": 2048,
"model_type": "chat"
},
{
"llm_name": "recursal/eagle-7b",
"tags": "LLM CHAT 9K",
"max_tokens": 10000,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-3.5-turbo-0613",
"tags": "LLM CHAT 4K",
"max_tokens": 4095,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-4-turbo-preview",
"tags": "LLM CHAT 125K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "undi95/remm-slerp-l2-13b:extended",
"tags": "LLM CHAT 6K",
"max_tokens": 6144,
"model_type": "chat"
},
{
"llm_name": "nousresearch/nous-hermes-2-mixtral-8x7b-dpo",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "nousresearch/nous-hermes-2-mixtral-8x7b-sft",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-tiny",
"tags": "LLM CHAT 32K",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-small",
"tags": "LLM CHAT 32K",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-medium",
"tags": "LLM CHAT 32K",
"max_tokens": 32000,
"model_type": "chat"
},
{
"llm_name": "austism/chronos-hermes-13b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "neversleep/noromaid-mixtral-8x7b-instruct",
"tags": "LLM CHAT 8K",
"max_tokens": 8000,
"model_type": "chat"
},
{
"llm_name": "nousresearch/nous-hermes-yi-34b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-7b-instruct-v0.2",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "cognitivecomputations/dolphin-mixtral-8x7b",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "google/gemini-pro",
"tags": "LLM CHAT 89K",
"max_tokens": 91728,
"model_type": "chat"
},
{
"llm_name": "google/gemini-pro-vision",
"tags": "LLM IMAGE2TEXT 44K",
"max_tokens": 45875,
"model_type": "image2text"
},
{
"llm_name": "mistralai/mixtral-8x7b",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "mistralai/mixtral-8x7b-instruct",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "rwkv/rwkv-5-world-3b",
"tags": "LLM CHAT 9K",
"max_tokens": 10000,
"model_type": "chat"
},
{
"llm_name": "recursal/rwkv-5-3b-ai-town",
"tags": "LLM CHAT 9K",
"max_tokens": 10000,
"model_type": "chat"
},
{
"llm_name": "togethercomputer/stripedhyena-nous-7b",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "togethercomputer/stripedhyena-hessian-7b",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "koboldai/psyfighter-13b-2",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "gryphe/mythomist-7b",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "openrouter/cinematika-7b",
"tags": "LLM CHAT 8K",
"max_tokens": 8000,
"model_type": "chat"
},
{
"llm_name": "nousresearch/nous-capybara-7b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "nousresearch/nous-capybara-7b:free",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "openchat/openchat-7b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "openchat/openchat-7b:free",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "neversleep/noromaid-20b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "gryphe/mythomist-7b:free",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "intel/neural-chat-7b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-2",
"tags": "LLM CHAT 195K",
"max_tokens": 200000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-2.1",
"tags": "LLM CHAT 195K",
"max_tokens": 200000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-instant-1.1",
"tags": "LLM CHAT 98K",
"max_tokens": 100000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-2:beta",
"tags": "LLM CHAT 195K",
"max_tokens": 200000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-2.1:beta",
"tags": "LLM CHAT 195K",
"max_tokens": 200000,
"model_type": "chat"
},
{
"llm_name": "teknium/openhermes-2.5-mistral-7b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "nousresearch/nous-capybara-34b",
"tags": "LLM CHAT 195K",
"max_tokens": 200000,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-4-vision-preview",
"tags": "LLM IMAGE2TEXT 125K",
"max_tokens": 128000,
"model_type": "image2text"
},
{
"llm_name": "lizpreciatior/lzlv-70b-fp16-hf",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "undi95/toppy-m-7b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "alpindale/goliath-120b",
"tags": "LLM CHAT 6K",
"max_tokens": 6144,
"model_type": "chat"
},
{
"llm_name": "undi95/toppy-m-7b:free",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "openrouter/auto",
"tags": "LLM CHAT 195K",
"max_tokens": 200000,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-3.5-turbo-1106",
"tags": "LLM CHAT 16K",
"max_tokens": 16385,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-4-1106-preview",
"tags": "LLM CHAT 125K",
"max_tokens": 128000,
"model_type": "chat"
},
{
"llm_name": "huggingfaceh4/zephyr-7b-beta:free",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "google/palm-2-chat-bison-32k",
"tags": "LLM CHAT 89K",
"max_tokens": 91750,
"model_type": "chat"
},
{
"llm_name": "google/palm-2-codechat-bison-32k",
"tags": "LLM CHAT 89K",
"max_tokens": 91750,
"model_type": "chat"
},
{
"llm_name": "teknium/openhermes-2-mistral-7b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "open-orca/mistral-7b-openorca",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "gryphe/mythomax-l2-13b:extended",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "xwin-lm/xwin-lm-70b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-3.5-turbo-instruct",
"tags": "LLM CHAT 4K",
"max_tokens": 4095,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-7b-instruct-v0.1",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-7b-instruct:free",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "pygmalionai/mythalion-13b",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-3.5-turbo-16k",
"tags": "LLM CHAT 16K",
"max_tokens": 16385,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-4-32k",
"tags": "LLM CHAT 32K",
"max_tokens": 32767,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-4-32k-0314",
"tags": "LLM CHAT 32K",
"max_tokens": 32767,
"model_type": "chat"
},
{
"llm_name": "meta-llama/codellama-34b-instruct",
"tags": "LLM CHAT 8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "phind/phind-codellama-34b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "nousresearch/nous-hermes-llama2-13b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "mancer/weaver",
"tags": "LLM CHAT 8K",
"max_tokens": 8000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-2.0",
"tags": "LLM CHAT 98K",
"max_tokens": 100000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-instant-1",
"tags": "LLM CHAT 98K",
"max_tokens": 100000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-1",
"tags": "LLM CHAT 98K",
"max_tokens": 100000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-1.2",
"tags": "LLM CHAT 98K",
"max_tokens": 100000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-instant-1.0",
"tags": "LLM CHAT 98K",
"max_tokens": 100000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-2.0:beta",
"tags": "LLM CHAT 98K",
"max_tokens": 100000,
"model_type": "chat"
},
{
"llm_name": "anthropic/claude-instant-1:beta",
"tags": "LLM CHAT 98K",
"max_tokens": 100000,
"model_type": "chat"
},
{
"llm_name": "undi95/remm-slerp-l2-13b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "google/palm-2-chat-bison",
"tags": "LLM CHAT 25K",
"max_tokens": 25804,
"model_type": "chat"
},
{
"llm_name": "google/palm-2-codechat-bison",
"tags": "LLM CHAT 19K",
"max_tokens": 20070,
"model_type": "chat"
},
{
"llm_name": "gryphe/mythomax-l2-13b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-2-13b-chat",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "meta-llama/llama-2-70b-chat",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-3.5-turbo",
"tags": "LLM CHAT 16K",
"max_tokens": 16385,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-3.5-turbo-0125",
"tags": "LLM CHAT 16K",
"max_tokens": 16385,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-3.5-turbo-0301",
"tags": "LLM CHAT 4K",
"max_tokens": 4095,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-4",
"tags": "LLM CHAT 8K",
"max_tokens": 8191,
"model_type": "chat"
},
{
"llm_name": "openai/gpt-4-0314",
"tags": "LLM CHAT 8K",
"max_tokens": 8191,
"model_type": "chat"
},
{
"llm_name": "01-ai/yi-large",
"tags": "LLM CHAT 32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "01-ai/yi-34b-200k",
"tags": "LLM CHAT 195K",
"max_tokens": 200000,
"model_type": "chat"
},
{
"llm_name": "01-ai/yi-34b-chat",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "01-ai/yi-34b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "01-ai/yi-6b",
"tags": "LLM CHAT 4K",
"max_tokens": 4096,
"model_type": "chat"
}
]
},
{
"name": "StepFun",
"logo": "",
"tags": "LLM",
"status": "1",
"llm": [
{
"llm_name": "step-1-8k",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "step-1-32k",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "step-1-128k",
"tags": "LLM,CHAT,128k",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "step-1-256k",
"tags": "LLM,CHAT,256k",
"max_tokens": 262144,
"model_type": "chat"
},
{
"llm_name": "step-1v-8k",
"tags": "LLM,CHAT,IMAGE2TEXT",
"max_tokens": 8192,
"model_type": "image2text"
}
]
},
{
"name": "NVIDIA",
"logo": "",
"tags": "LLM,TEXT EMBEDDING, TEXT RE-RANK",
"status": "1",
"llm": [
{
"llm_name": "nvidia/nemotron-4-340b-reward",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "aisingapore/sea-lion-7b-instruct",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "databricks/dbrx-instruct",
"tags": "LLM,CHAT,16K",
"max_tokens": 16384,
"model_type": "chat"
},
{
"llm_name": "google/gemma-7b",
"tags": "LLM,CHAT,32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "google/gemma-2b",
"tags": "LLM,CHAT,16K",
"max_tokens": 16384,
"model_type": "chat"
},
{
"llm_name": "google/gemma-2-9b-it",
"tags": "LLM,CHAT,8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "google/gemma-2-27b-it",
"tags": "LLM,CHAT,8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "google/recurrentgemma-2b",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "mediatek/breeze-7b-instruct",
"tags": "LLM,CHAT,8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "meta/llama2-70b",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "meta/llama3-8b",
"tags": "LLM,CHAT,8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "meta/llama3-70b",
"tags": "LLM,CHAT,8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "microsoft/phi-3-medium-128k-instruct",
"tags": "LLM,CHAT,128K",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "microsoft/phi-3-medium-4k-instruct",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "microsoftphi-3-mini-128k-instruct",
"tags": "LLM,CHAT,128K",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "microsoft/phi-3-mini-4k-instruct",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "microsoft/phi-3-small-128k-instruct",
"tags": "LLM,CHAT,128K",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "microsoft/phi-3-small-8k-instruct",
"tags": "LLM,CHAT,8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-7b-instruct",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-7b-instruct-v0.3",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "mistralai/mixtral-8x7b-instruct",
"tags": "LLM,CHAT,32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "mistralai/mixtral-8x22b-instruct",
"tags": "LLM,CHAT,64K",
"max_tokens": 65536,
"model_type": "chat"
},
{
"llm_name": "mistralai/mistral-large",
"tags": "LLM,CHAT,32K",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "nv-mistralai/mistral-nemo-12b-instruct",
"tags": "LLM,CHAT,128K",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "nvidia/llama3-chatqa-1.5-70b",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "nvidia/llama3-chatqa-1.5-8b",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "nvidia/nemotron-4-340b-instruct",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "seallms/seallm-7b-v2.5",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "snowflake/arctic",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "upstage/solar-10.7b-instruct",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "baai/bge-m3",
"tags": "TEXT EMBEDDING,8K",
"max_tokens": 8192,
"model_type": "embedding"
},
{
"llm_name": "nvidia/embed-qa-4",
"tags": "TEXT EMBEDDING,512",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "nvidia/nv-embed-v1",
"tags": "TEXT EMBEDDING,32K",
"max_tokens": 32768,
"model_type": "embedding"
},
{
"llm_name": "nvidia/nv-embedqa-e5-v5",
"tags": "TEXT EMBEDDING,512",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "nvidia/nv-embedqa-mistral-7b-v2",
"tags": "TEXT EMBEDDING,512",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "nvidia/nv-rerankqa-mistral-4b-v3",
"tags": "RE-RANK,512",
"max_tokens": 512,
"model_type": "rerank"
},
{
"llm_name": "nvidia/rerank-qa-mistral-4b",
"tags": "RE-RANK,512",
"max_tokens": 512,
"model_type": "rerank"
},
{
"llm_name": "snowflake/arctic-embed-l",
"tags": "TEXT EMBEDDING,512",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "adept/fuyu-8b",
"tags": "LLM,IMAGE2TEXT,4K",
"max_tokens": 4096,
"model_type": "image2text"
},
{
"llm_name": "google/deplot",
"tags": "LLM,IMAGE2TEXT,4K",
"max_tokens": 4096,
"model_type": "image2text"
},
{
"llm_name": "google/paligemma",
"tags": "LLM,IMAGE2TEXT,4K",
"max_tokens": 4096,
"model_type": "image2text"
},
{
"llm_name": "Iiuhaotian/Ilava-v1.6-34b",
"tags": "LLM,IMAGE2TEXT,4K",
"max_tokens": 4096,
"model_type": "image2text"
},
{
"llm_name": "Iiuhaotian/Ilava-v1.6-mistral-7b",
"tags": "LLM,IMAGE2TEXT,4K",
"max_tokens": 4096,
"model_type": "image2text"
},
{
"llm_name": "microsoft/kosmos-2",
"tags": "LLM,IMAGE2TEXT,4K",
"max_tokens": 4096,
"model_type": "image2text"
},
{
"llm_name": "microsoft/phi-3-vision-128k-instruct",
"tags": "LLM,IMAGE2TEXT,128K",
"max_tokens": 131072,
"model_type": "image2text"
},
{
"llm_name": "nvidia/neva-22b",
"tags": "LLM,IMAGE2TEXT,4K",
"max_tokens": 4096,
"model_type": "image2text"
}
]
},
{
"name": "LM-Studio",
"logo": "",
"tags": "LLM,TEXT EMBEDDING,IMAGE2TEXT",
"status": "1",
"llm": []
},
{
"name": "cohere",
"logo": "",
"tags": "LLM,TEXT EMBEDDING, TEXT RE-RANK",
"status": "1",
"llm": [
{
"llm_name": "command-r-plus",
"tags": "LLM,CHAT,128k",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "command-r",
"tags": "LLM,CHAT,128k",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "command",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "command-nightly",
"tags": "LLM,CHAT,128k",
"max_tokens": 131072,
"model_type": "chat"
},
{
"llm_name": "command-light",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "command-light-nightly",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "embed-english-v3.0",
"tags": "TEXT EMBEDDING",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "embed-english-light-v3.0",
"tags": "TEXT EMBEDDING",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "embed-multilingual-v3.0",
"tags": "TEXT EMBEDDING",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "embed-multilingual-light-v3.0",
"tags": "TEXT EMBEDDING",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "embed-english-v2.0",
"tags": "TEXT EMBEDDING",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "embed-english-light-v2.0",
"tags": "TEXT EMBEDDING",
"max_tokens": 512,
"model_type": "embedding"
},
{
"llm_name": "embed-multilingual-v2.0",
"tags": "TEXT EMBEDDING",
"max_tokens": 256,
"model_type": "embedding"
},
{
"llm_name": "rerank-english-v3.0",
"tags": "RE-RANK,4k",
"max_tokens": 4096,
"model_type": "rerank"
},
{
"llm_name": "rerank-multilingual-v3.0",
"tags": "RE-RANK,4k",
"max_tokens": 4096,
"model_type": "rerank"
},
{
"llm_name": "rerank-english-v2.0",
"tags": "RE-RANK,512",
"max_tokens": 8196,
"model_type": "rerank"
},
{
"llm_name": "rerank-multilingual-v2.0",
"tags": "RE-RANK,512",
"max_tokens": 512,
"model_type": "rerank"
}
]
},
{
"name": "Lepton",
"logo": "",
"tags": "LLM",
"status": "1",
"llm": [
{
"llm_name": "dolphin-mixtral-8x7b",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "gemma-7b",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "llama3-1-8b",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "llama3-8b",
"tags": "LLM,CHAT,8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "llama2-13b",
"tags": "LLM,CHAT,4K",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "llama3-1-70b",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "llama3-70b",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "llama3-1-405b",
"tags": "LLM,CHAT,8k",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "mistral-7b",
"tags": "LLM,CHAT,8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "mistral-8x7b",
"tags": "LLM,CHAT,8K",
"max_tokens": 8192,
"model_type": "chat"
},
{
"llm_name": "nous-hermes-llama2",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "openchat-3-5",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "toppy-m-7b",
"tags": "LLM,CHAT,4k",
"max_tokens": 4096,
"model_type": "chat"
},
{
"llm_name": "wizardlm-2-7b",
"tags": "LLM,CHAT,32k",
"max_tokens": 32768,
"model_type": "chat"
},
{
"llm_name": "wizardlm-2-8x22b",
"tags": "LLM,CHAT,64K",
"max_tokens": 65536,
"model_type": "chat"
}
]
}
]
}