# # model_download.py
# from modelscope import snapshot_download
# model_dir = snapshot_download('qwen/Qwen2.5-32B-Instruct', cache_dir='/root/autodl-fs', revision='master')

# #模型下载
# from modelscope import snapshot_download
# model_dir = snapshot_download('Qwen/Qwen2.5-32B-Instruct-GPTQ-Int4', cache_dir='/root/autodl-fs')

#模型下载
from modelscope import snapshot_download
model_dir = snapshot_download('Qwen/Qwen2.5-14B-Instruct', cache_dir='/root/autodl-fs')

# python -m vllm.entrypoints.openai.api_server --model /root/autodl-fs/Qwen/Qwen2___5-14B-Instruct  --served-model-name Qwen2.5-14B-Instruct --max-model-len=2000 --port 8000

# python -m vllm.entrypoints.openai.api_server --model /root/autodl-fs/Qwen/Qwen2___5-32B-Instruct-GPTQ-Int4  --served-model-name Qwen2.5-32B-Instruct-GPTQ-Int4 --max-model-len=2048 --port 8000