VictorSanh commited on
Commit
042af7d
1 Parent(s): eea1236

update readme

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -242,7 +242,7 @@ Flash attention 2 support is available both for `idefics2-8b-base` and `idefics2
242
 
243
  <details><summary>Click to expand.</summary>
244
 
245
- 4-bit AWQ-quantized versions of the checkpoints are also available and allow module fusing for accelerated inference. First make sure you install the Auto-AWQ library with `pip install autoawq`.
246
 
247
  ```diff
248
  + from transformers import AwqConfig
 
242
 
243
  <details><summary>Click to expand.</summary>
244
 
245
+ 4-bit AWQ-quantized versions of the checkpoints are also available and allow module fusing for accelerated inference. First make sure you install the Auto-AWQ library with `pip install autoawq`. Also make sure that this [fix] is integrated into your installation.
246
 
247
  ```diff
248
  + from transformers import AwqConfig