from modelscope import snapshot_download
model_dir = snapshot_download('deepseek-ai/DeepSeek-R1-Distill-Qwen-7B', cache_dir='/root/autodl-tmp/model')


# python -m vllm.entrypoints.openai.api_server --model /root/autodl-tmp/model/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B  --served-model-name DeepSeek-R1-Distill-Qwen-7B-01 --max-model-len=2048 --tensor-parallel-size 1 --gpu-memory-utilization 0.45 --port 8000

# python -m vllm.entrypoints.openai.api_server --model /root/autodl-tmp/model/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B  --served-model-name DeepSeek-R1-Distill-Qwen-7B-02 --max-model-len=2048 --tensor-parallel-size 1 --gpu-memory-utilization 0.45 --port 8001