KinglyCrow
commited on
Commit
•
1f0403c
1
Parent(s):
64be886
Update README.md
Browse files
README.md
CHANGED
@@ -8,6 +8,8 @@ Make sure to run with FlashAttention like in https://github.com/huggingface/text
|
|
8 |
|
9 |
Also note the GPTQ 4bit quantized version seems to run about 2x slower compared to the 8bit bitsandbytes version within text-generation-inference, typically we were seeing about 600-800ms latency for token generation for 8bit bitsandbytes whereas we're seeing about 1.2-1.7s with the 4bit GPTQ version.
|
10 |
|
|
|
|
|
11 |
This was quantized using:
|
12 |
|
13 |
`text-generation-server quantize tiiuae/falcon-40b-instruct /tmp/falcon40instructgptq --upload-to-model-id AxisMind/falcon-40b-instruct-gptq --trust-remote-code --act-order`
|
|
|
8 |
|
9 |
Also note the GPTQ 4bit quantized version seems to run about 2x slower compared to the 8bit bitsandbytes version within text-generation-inference, typically we were seeing about 600-800ms latency for token generation for 8bit bitsandbytes whereas we're seeing about 1.2-1.7s with the 4bit GPTQ version.
|
10 |
|
11 |
+
VRAM usage is a little over 25gb for this 4bit quantized version, compared to 47gb for the 8bit and 80gb for full.
|
12 |
+
|
13 |
This was quantized using:
|
14 |
|
15 |
`text-generation-server quantize tiiuae/falcon-40b-instruct /tmp/falcon40instructgptq --upload-to-model-id AxisMind/falcon-40b-instruct-gptq --trust-remote-code --act-order`
|