bartowski commited on
Commit
a0cbdba
·
verified ·
1 Parent(s): 4090c97

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -47,6 +47,7 @@ Run them in [LM Studio](https://lmstudio.ai/)
47
  | [gemma-2-9b-it-abliterated-Q4_0_8_8.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q4_0_8_8.gguf) | Q4_0_8_8 | 5.44GB | false | Optimized for ARM inference. Requires 'sve' support (see link below). |
48
  | [gemma-2-9b-it-abliterated-Q4_0_4_8.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q4_0_4_8.gguf) | Q4_0_4_8 | 5.44GB | false | Optimized for ARM inference. Requires 'i8mm' support (see link below). |
49
  | [gemma-2-9b-it-abliterated-Q4_0_4_4.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q4_0_4_4.gguf) | Q4_0_4_4 | 5.44GB | false | Optimized for ARM inference. Should work well on all ARM chips, pick this if you're unsure. |
 
50
  | [gemma-2-9b-it-abliterated-Q3_K_XL.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q3_K_XL.gguf) | Q3_K_XL | 5.35GB | false | Uses Q8_0 for embed and output weights. Lower quality but usable, good for low RAM availability. |
51
  | [gemma-2-9b-it-abliterated-IQ4_XS.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-IQ4_XS.gguf) | IQ4_XS | 5.18GB | false | Decent quality, smaller than Q4_K_S with similar performance, *recommended*. |
52
  | [gemma-2-9b-it-abliterated-Q3_K_L.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q3_K_L.gguf) | Q3_K_L | 5.13GB | false | Lower quality but usable, good for low RAM availability. |
 
47
  | [gemma-2-9b-it-abliterated-Q4_0_8_8.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q4_0_8_8.gguf) | Q4_0_8_8 | 5.44GB | false | Optimized for ARM inference. Requires 'sve' support (see link below). |
48
  | [gemma-2-9b-it-abliterated-Q4_0_4_8.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q4_0_4_8.gguf) | Q4_0_4_8 | 5.44GB | false | Optimized for ARM inference. Requires 'i8mm' support (see link below). |
49
  | [gemma-2-9b-it-abliterated-Q4_0_4_4.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q4_0_4_4.gguf) | Q4_0_4_4 | 5.44GB | false | Optimized for ARM inference. Should work well on all ARM chips, pick this if you're unsure. |
50
+ | [gemma-2-9b-it-abliterated-Q4_0.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q4_0.gguf) | Q4_0 | 5.44GB | false | Legacy format, offers online repacking for ARM and AVX inference. |
51
  | [gemma-2-9b-it-abliterated-Q3_K_XL.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q3_K_XL.gguf) | Q3_K_XL | 5.35GB | false | Uses Q8_0 for embed and output weights. Lower quality but usable, good for low RAM availability. |
52
  | [gemma-2-9b-it-abliterated-IQ4_XS.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-IQ4_XS.gguf) | IQ4_XS | 5.18GB | false | Decent quality, smaller than Q4_K_S with similar performance, *recommended*. |
53
  | [gemma-2-9b-it-abliterated-Q3_K_L.gguf](https://huggingface.co/bartowski/gemma-2-9b-it-abliterated-GGUF/blob/main/gemma-2-9b-it-abliterated-Q3_K_L.gguf) | Q3_K_L | 5.13GB | false | Lower quality but usable, good for low RAM availability. |