Update README.md
Browse files
README.md
CHANGED
@@ -6,4 +6,6 @@ Quant of https://huggingface.co/junelee/wizard-vicuna-13b tested working with Oc
|
|
6 |
|
7 |
Someone made a Triton quant already here, but it will not work with Occam's KoboldAI/GPTQ fork: https://huggingface.co/fbjr/wizard-vicuna-13b-4bit-128g
|
8 |
|
|
|
|
|
9 |
```python llama.py ./wizard-vicuna-13b c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors 4bit-128g.safetensors```
|
|
|
6 |
|
7 |
Someone made a Triton quant already here, but it will not work with Occam's KoboldAI/GPTQ fork: https://huggingface.co/fbjr/wizard-vicuna-13b-4bit-128g
|
8 |
|
9 |
+
Note that this model is fairly heavily censored (in my opinion) and delivers AI-moralizing responses to prompts that Vicuna 1.1 does not complain about.
|
10 |
+
|
11 |
```python llama.py ./wizard-vicuna-13b c4 --wbits 4 --true-sequential --groupsize 128 --save_safetensors 4bit-128g.safetensors```
|