keyfan commited on
Commit
ccc3bf0
1 Parent(s): d01d282

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -9,6 +9,8 @@ This is 2-bit quantization of [mistralai/Mixtral-8x7B-Instruct-v0.1](https://hug
9
  ## Model loading
10
  Please follow the instruction of [QuIP-for-all](https://github.com/chu-tianxiang/QuIP-for-all) for usage.
11
 
 
 
12
  ## Perplexity
13
  Measured at Wikitext with 4096 context length
14
  | fp16 | 2-bit |
@@ -16,4 +18,5 @@ Measured at Wikitext with 4096 context length
16
  | 3.8825 | 5.2799 |
17
 
18
  ## Speed
19
- About 16 tokens/s at batchsize = 1 using vLLM at single A100.
 
 
9
  ## Model loading
10
  Please follow the instruction of [QuIP-for-all](https://github.com/chu-tianxiang/QuIP-for-all) for usage.
11
 
12
+ As an alternative, you can use [vLLM branch](https://github.com/chu-tianxiang/vllm-gptq/tree/quip_gemv) for faster inference. QuIP has to launch like 5 kernels for each linear layer, so it's very helpful for vLLM to use cuda-graph to reduce launching overhead. BTW, If you have problem installing fast-hadamard-transform from pip, you can also install it from [source](https://github.com/Dao-AILab/fast-hadamard-transform)
13
+
14
  ## Perplexity
15
  Measured at Wikitext with 4096 context length
16
  | fp16 | 2-bit |
 
18
  | 3.8825 | 5.2799 |
19
 
20
  ## Speed
21
+ Measured with `examples/benchmark_latency.py` script at vLLM repo.
22
+ At batch size = 1, it generates at 16.3 tokens/s with single 3090.