Text Generation
Transformers
Safetensors
minimax_m2
conversational
custom_code
fp8
oql commited on
Commit
f7804c9
·
verified ·
1 Parent(s): 757303d

Add KTransformers Support in README.md

Browse files

Add KTransformers Support in README.md for MiniMax-M2.1.

Files changed (1) hide show
  1. README.md +4 -0
README.md CHANGED
@@ -184,6 +184,10 @@ We recommend using [SGLang](https://docs.sglang.ai/) to serve MiniMax-M2. SGLang
184
 
185
  We recommend using [vLLM](https://docs.vllm.ai/en/stable/) to serve MiniMax-M2. vLLM provides efficient day-0 support of MiniMax-M2 model, check https://docs.vllm.ai/projects/recipes/en/latest/MiniMax/MiniMax-M2.html for latest deployment guide. We also provide our [vLLM Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/vllm_deploy_guide.md).
186
 
 
 
 
 
187
  ### MLX
188
 
189
  We recommend using [MLX-LM](https://github.com/ml-explore/mlx-lm) to serve MiniMax-M2. Please refer to our [MLX Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/mlx_deploy_guide.md) for more details.
 
184
 
185
  We recommend using [vLLM](https://docs.vllm.ai/en/stable/) to serve MiniMax-M2. vLLM provides efficient day-0 support of MiniMax-M2 model, check https://docs.vllm.ai/projects/recipes/en/latest/MiniMax/MiniMax-M2.html for latest deployment guide. We also provide our [vLLM Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/vllm_deploy_guide.md).
186
 
187
+ ### KTransformers
188
+
189
+ We recommend using [KTransformers](https://github.com/kvcache-ai/ktransformers) to serve MiniMax-M2.1. KTransformers provides efficient day-0 support for MiniMax-M2.1 model and can run the native weights with **≥32GB VRAM** and **≥256GB DRAM**. For installation and usage, see [KTransformers MiniMax-M2.1 Tutorial](https://github.com/kvcache-ai/ktransformers/blob/main/doc/en/MiniMax-M2.1-Tutorial.md).
190
+
191
  ### MLX
192
 
193
  We recommend using [MLX-LM](https://github.com/ml-explore/mlx-lm) to serve MiniMax-M2. Please refer to our [MLX Deployment Guide](https://huggingface.co/MiniMaxAI/MiniMax-M2/blob/main/docs/mlx_deploy_guide.md) for more details.