Update README.md
Browse files
README.md
CHANGED
@@ -200,7 +200,7 @@ It was created with group_size none (-1) to reduce VRAM usage, and with --act-or
|
|
200 |
* `gptq_model-4bit-128g.safetensors`
|
201 |
* Works with AutoGPTQ in CUDA or Triton modes.
|
202 |
* Does NOT work with [ExLlama](https://github.com/turboderp/exllama) as it's not a Llama model.
|
203 |
-
*
|
204 |
* Works with text-generation-webui, including one-click-installers.
|
205 |
* Parameters: Groupsize = -1. Act Order / desc_act = True.
|
206 |
|
|
|
200 |
* `gptq_model-4bit-128g.safetensors`
|
201 |
* Works with AutoGPTQ in CUDA or Triton modes.
|
202 |
* Does NOT work with [ExLlama](https://github.com/turboderp/exllama) as it's not a Llama model.
|
203 |
+
* Untested with GPTQ-for-LLaMa.
|
204 |
* Works with text-generation-webui, including one-click-installers.
|
205 |
* Parameters: Groupsize = -1. Act Order / desc_act = True.
|
206 |
|