
conda activate llm_knowledge & python -m fastchat.serve.controller --host 0.0.0.0 &> ./log/controller.log  &
conda activate llm_knowledge & python -m fastchat.serve.model_worker --model-path /home/cat/Documents/Project/llm_knowledge/models/qwen/Qwen1___5-0___5B-Chat/ --host 0.0.0.0 &> ./log/model_worker.log &
conda activate llm_knowledge & python -m fastchat.serve.openai_api_server --host 0.0.0.0 &> ./log/openai_api_server.log &
