JosephusCheung
commited on
Commit
•
7331bed
1
Parent(s):
1750b03
Update README.md
Browse files
README.md
CHANGED
@@ -22,7 +22,7 @@ tags:
|
|
22 |
|
23 |
Free T4 Colab demo, please check 4bit version: [JosephusCheung/GuanacoOnConsumerHardware](https://huggingface.co/JosephusCheung/GuanacoOnConsumerHardware).**
|
24 |
|
25 |
-
**It is highly recommended to use fp16 inference for this model, as 8-bit precision may significantly affect performance. If you require a more Consumer Hardware friendly version, please use the specialized quantized [JosephusCheung/GuanacoOnConsumerHardware](https://huggingface.co/JosephusCheung/GuanacoOnConsumerHardware)
|
26 |
|
27 |
**You are encouraged to use the latest version of transformers from GitHub.**
|
28 |
|
|
|
22 |
|
23 |
Free T4 Colab demo, please check 4bit version: [JosephusCheung/GuanacoOnConsumerHardware](https://huggingface.co/JosephusCheung/GuanacoOnConsumerHardware).**
|
24 |
|
25 |
+
**It is highly recommended to use fp16 inference for this model, as 8-bit precision may significantly affect performance. If you require a more Consumer Hardware friendly version, please use the specialized quantized, only 5+GB V-Ram required** [JosephusCheung/GuanacoOnConsumerHardware](https://huggingface.co/JosephusCheung/GuanacoOnConsumerHardware).
|
26 |
|
27 |
**You are encouraged to use the latest version of transformers from GitHub.**
|
28 |
|