---
title: Customizing Provider Model List in LobeChat for Deployment
description: >-
  Learn how to customize the model list in LobeChat for deployment with the syntax and extension capabilities

tags:
  - LobeChat
  - model customization
  - deployment
  - extension capabilities
---

# Model List

LobeChat supports customizing the model list during deployment. This configuration is done in the environment for each [model provider](/docs/self-hosting/environment-variables/model-provider).

You can use `+` to add a model, `-` to hide a model, and use `model name->deploymentName=display name<extension configuration>` to customize the display name of a model, separated by English commas. The basic syntax is as follows:

```text
id->deploymentName=displayName<maxToken:vision:reasoning:search:fc:file:imageOutput>,model2,model3
```

The deploymentName `->deploymentName` can be omitted, and it defaults to the latest model version. Currently, the model service providers that support `->deploymentName` are: Azure, Volcengine and Qwen.

For example: `+qwen-7b-chat,+glm-6b,-gpt-3.5-turbo,gpt-4-0125-preview=gpt-4-turbo`

In the above example, it adds `qwen-7b-chat` and `glm-6b` to the model list, removes `gpt-3.5-turbo` from the list, and displays the model name of `gpt-4-0125-preview` as `gpt-4-turbo`. If you want to disable all models first and then enable specific models, you can use `-all,+gpt-3.5-turbo`, which means only enabling `gpt-3.5-turbo`.

### -all: Hide all models

- Description: `-all` means hiding all built-in models first. It’s usually combined with `+` to only enable the models you explicitly specify.
- Example:

```text
-all,+gpt-3.5-turbo,+gpt-4-0125-preview=gpt-4-turbo
```

This enables only gpt-3.5-turbo and gpt-4-turbo while hiding other models.

## Extension Capabilities

Considering the diversity of model capabilities, we started to add extension configuration in version `0.147.8`, with the following rules:

```shell
id->deploymentName=displayName<maxToken:vision:reasoning:search:fc:file:imageOutput>
```

The first value in angle brackets is designated as the `maxToken` for this model. The second value and beyond are the model's extension capabilities, separated by colons `:`, and the order is not important.

Examples are as follows:

- `chatglm-6b=ChatGLM 6B<4096>`: ChatGLM 6B, maximum context of 4k, no advanced capabilities;
- `spark-v3.5=讯飞星火 v3.5<8192:fc>`: Xunfei Spark 3.5 model, maximum context of 8k, supports Function Call;
- `gemini-1.5-flash-latest=Gemini 1.5 Flash<16000:vision>`: Google Vision model, maximum context of 16k, supports image recognition;
- `o3-mini=OpenAI o3-mini<200000:reasoning:fc>`: OpenAI o3-mini model, maximum context of 200k, supports reasoning and Function Call;
- `qwen-max-latest=Qwen Max<32768:search:fc>`: Qwen 2.5 Max model, maximum context of 32k, supports web search and Function Call;
- `gpt-4-all=ChatGPT Plus<128000:fc:vision:file>`, hacked version of ChatGPT Plus web, context of 128k, supports image recognition, Function Call, file upload;
- `gemini-2.0-flash-exp-image-generation=Gemini 2.0 Flash (Image Generation) Experimental<32768:imageOutput:vision>`, Gemini 2.0 Flash Experimental model for image generation, maximum context of 32k, supports image generation and recognition.

Currently supported extension capabilities are:

| ---           | Description                                              |
| ------------- | -------------------------------------------------------- |
| `fc`          | Function Calling                                         |
| `vision`      | Image Recognition                                        |
| `imageOutput` | Image Generation                                         |
| `reasoning`   | Support Reasoning                                        |
| `search`      | Support Web Search                                       |
| `file`        | File Upload (a bit hacky, not recommended for daily use) |
