Text Generation
Transformers
Safetensors
English
llama
text-generation-inference
4-bit precision
gptq
TheBloke commited on
Commit
31226c5
·
1 Parent(s): 37176e6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -38,7 +38,7 @@ Open the text-generation-webui UI as normal.
38
 
39
  This will work with all versions of GPTQ-for-LLaMa. It has maximum compatibility.
40
 
41
- It was created without `--act-order` to increase quantisation quality, but without group_size so as to reduce VRAM requirements.
42
 
43
  * `wizard-mega-13B-GPTQ-4bit-128g.safetensors`
44
  * Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
@@ -46,7 +46,7 @@ It was created without `--act-order` to increase quantisation quality, but witho
46
  * Parameters: Groupsize = 128. No act-order.
47
  * Command used to create the GPTQ:
48
  ```
49
- python llama.py /workspace/models/openaccess-ai-collective_wizard-mega-13b c4 --wbits 4 --act-order --true-sequential --groupsize 128 --save_safetensors /workspace/wizard-mega-13b/gptq/wizard-mega-13B-GPTQ-4bit-128g.no-act-order.safetensors
50
  ```
51
 
52
 
 
38
 
39
  This will work with all versions of GPTQ-for-LLaMa. It has maximum compatibility.
40
 
41
+ It was created without `--act-order` to ensure compatibility with all UIs out there.
42
 
43
  * `wizard-mega-13B-GPTQ-4bit-128g.safetensors`
44
  * Works with all versions of GPTQ-for-LLaMa code, both Triton and CUDA branches
 
46
  * Parameters: Groupsize = 128. No act-order.
47
  * Command used to create the GPTQ:
48
  ```
49
+ python llama.py /workspace/models/openaccess-ai-collective_wizard-mega-13b c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors /workspace/wizard-mega-13b/gptq/wizard-mega-13B-GPTQ-4bit-128g.no-act-order.safetensors
50
  ```
51
 
52