FrankC0st1e
commited on
Commit
•
9c8e6ee
1
Parent(s):
e00ba39
Update README.md
Browse files
README.md
CHANGED
@@ -22,7 +22,7 @@ print(responds)
|
|
22 |
```
|
23 |
|
24 |
# Note
|
25 |
-
1. You can alse inference with [vLLM](https://github.com/vllm-project/vllm), which
|
26 |
2. The precision of model weights in this repo is bfloat16. Manual convertion is needed for other kinds of dtype.
|
27 |
3. For more details, please refer to our [github repo](https://github.com/OpenBMB/MiniCPM).
|
28 |
|
|
|
22 |
```
|
23 |
|
24 |
# Note
|
25 |
+
1. You can alse inference with [vLLM](https://github.com/vllm-project/vllm)(>=0.4.1), which is compatible with this repo and has a much higher inference throughput.
|
26 |
2. The precision of model weights in this repo is bfloat16. Manual convertion is needed for other kinds of dtype.
|
27 |
3. For more details, please refer to our [github repo](https://github.com/OpenBMB/MiniCPM).
|
28 |
|