RangiLyu commited on
Commit
292bc88
1 Parent(s): f470045

update readme

Browse files
Files changed (1) hide show
  1. README.md +32 -0
README.md CHANGED
@@ -125,6 +125,8 @@ Or you can launch an OpenAI compatible server with the following command:
125
  lmdeploy serve api_server internlm/internlm2-chat-7b --model-name internlm2-chat-7b --server-port 23333
126
  ```
127
 
 
 
128
  ```bash
129
  curl http://localhost:23333/v1/chat/completions \
130
  -H "Content-Type: application/json" \
@@ -151,6 +153,8 @@ pip install vllm
151
  python -m vllm.entrypoints.openai.api_server --model internlm/internlm2-chat-7b --served-model-name internlm2-chat-7b --trust-remote-code
152
  ```
153
 
 
 
154
  ```bash
155
  curl http://localhost:8000/v1/chat/completions \
156
  -H "Content-Type: application/json" \
@@ -272,6 +276,20 @@ print(response)
272
  lmdeploy serve api_server internlm/internlm2-chat-7b --server-port 23333
273
  ```
274
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
275
  更多信息请查看 [LMDeploy 文档](https://lmdeploy.readthedocs.io/en/latest/)
276
 
277
  ### vLLM
@@ -286,6 +304,20 @@ pip install vllm
286
  python -m vllm.entrypoints.openai.api_server --model internlm/internlm2-chat-7b --trust-remote-code
287
  ```
288
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
289
  更多信息请查看 [vLLM 文档](https://docs.vllm.ai/en/latest/index.html)
290
 
291
  ## 开源许可证
 
125
  lmdeploy serve api_server internlm/internlm2-chat-7b --model-name internlm2-chat-7b --server-port 23333
126
  ```
127
 
128
+ Then you can send a chat request to the server:
129
+
130
  ```bash
131
  curl http://localhost:23333/v1/chat/completions \
132
  -H "Content-Type: application/json" \
 
153
  python -m vllm.entrypoints.openai.api_server --model internlm/internlm2-chat-7b --served-model-name internlm2-chat-7b --trust-remote-code
154
  ```
155
 
156
+ Then you can send a chat request to the server:
157
+
158
  ```bash
159
  curl http://localhost:8000/v1/chat/completions \
160
  -H "Content-Type: application/json" \
 
276
  lmdeploy serve api_server internlm/internlm2-chat-7b --server-port 23333
277
  ```
278
 
279
+ 然后你可以向服务端发起一个聊天请求:
280
+
281
+ ```bash
282
+ curl http://localhost:23333/v1/chat/completions \
283
+ -H "Content-Type: application/json" \
284
+ -d '{
285
+ "model": "internlm2-chat-7b",
286
+ "messages": [
287
+ {"role": "system", "content": "You are a helpful assistant."},
288
+ {"role": "user", "content": "Introduce deep learning to me."}
289
+ ]
290
+ }'
291
+ ```
292
+
293
  更多信息请查看 [LMDeploy 文档](https://lmdeploy.readthedocs.io/en/latest/)
294
 
295
  ### vLLM
 
304
  python -m vllm.entrypoints.openai.api_server --model internlm/internlm2-chat-7b --trust-remote-code
305
  ```
306
 
307
+ 然后你可以向服务端发起一个聊天请求:
308
+
309
+ ```bash
310
+ curl http://localhost:8000/v1/chat/completions \
311
+ -H "Content-Type: application/json" \
312
+ -d '{
313
+ "model": "internlm2-chat-7b",
314
+ "messages": [
315
+ {"role": "system", "content": "You are a helpful assistant."},
316
+ {"role": "user", "content": "Introduce deep learning to me."}
317
+ ]
318
+ }'
319
+ ```
320
+
321
  更多信息请查看 [vLLM 文档](https://docs.vllm.ai/en/latest/index.html)
322
 
323
  ## 开源许可证