Bojun-Feng
commited on
Commit
•
d8d84d4
1
Parent(s):
199529e
Update README.md
Browse files
README.md
CHANGED
@@ -18,7 +18,9 @@ GGML files are for CPU + GPU inference using [chatglm.cpp](https://github.com/li
|
|
18 |
| chatglm2-ggml-q4_0.bin | q4_0 | 4 | 3.5 GB |
|
19 |
| chatglm2-ggml-q4_1.bin | q4_1 | 4 | 3.9 GB |
|
20 |
| chatglm2-ggml-q5_0.bin | q5_0 | 5 | 4.3 GB |
|
21 |
-
| chatglm2-ggml-q5_1.bin | q5_1 |
|
|
|
|
|
22 |
|
23 |
# How to run in xorbits-inference
|
24 |
Coming soon.
|
|
|
18 |
| chatglm2-ggml-q4_0.bin | q4_0 | 4 | 3.5 GB |
|
19 |
| chatglm2-ggml-q4_1.bin | q4_1 | 4 | 3.9 GB |
|
20 |
| chatglm2-ggml-q5_0.bin | q5_0 | 5 | 4.3 GB |
|
21 |
+
| chatglm2-ggml-q5_1.bin | q5_1 | 5 | 4.7 GB |
|
22 |
+
| chatglm2-ggml-q5_1.bin | q8_0 | 8 | 6.6 GB |
|
23 |
+
|
24 |
|
25 |
# How to run in xorbits-inference
|
26 |
Coming soon.
|