VictorSanh
commited on
Commit
•
042af7d
1
Parent(s):
eea1236
update readme
Browse files
README.md
CHANGED
@@ -242,7 +242,7 @@ Flash attention 2 support is available both for `idefics2-8b-base` and `idefics2
|
|
242 |
|
243 |
<details><summary>Click to expand.</summary>
|
244 |
|
245 |
-
4-bit AWQ-quantized versions of the checkpoints are also available and allow module fusing for accelerated inference. First make sure you install the Auto-AWQ library with `pip install autoawq`.
|
246 |
|
247 |
```diff
|
248 |
+ from transformers import AwqConfig
|
|
|
242 |
|
243 |
<details><summary>Click to expand.</summary>
|
244 |
|
245 |
+
4-bit AWQ-quantized versions of the checkpoints are also available and allow module fusing for accelerated inference. First make sure you install the Auto-AWQ library with `pip install autoawq`. Also make sure that this [fix] is integrated into your installation.
|
246 |
|
247 |
```diff
|
248 |
+ from transformers import AwqConfig
|