Update README.md
Browse files
README.md
CHANGED
@@ -37,19 +37,15 @@ Fixing tokenizer
|
|
37 |
| -------- | ---------- | --------- | ----- | ----------- |
|
38 |
| [Yi-Coder-9B-Chat-f16.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-f16.gguf) | f16 | 17.66GB | false | Full F16 weights. |
|
39 |
| [Yi-Coder-9B-Chat-Q8_0.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q8_0.gguf) | Q8_0 | 9.38GB | false | Extremely high quality, generally unneeded but max available quant. |
|
40 |
-
| [Yi-Coder-9B-Chat-Q8_0.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q8_0.gguf) | Q8_0 | 9.38GB | false | Extremely high quality, generally unneeded but max available quant. |
|
41 |
| [Yi-Coder-9B-Chat-Q6_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q6_K_L.gguf) | Q6_K_L | 7.37GB | false | Uses Q8_0 for embed and output weights. Very high quality, near perfect, *recommended*. |
|
42 |
| [Yi-Coder-9B-Chat-Q6_K.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q6_K.gguf) | Q6_K | 7.25GB | false | Very high quality, near perfect, *recommended*. |
|
43 |
-
| [Yi-Coder-9B-Chat-Q6_K.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q6_K.gguf) | Q6_K | 7.25GB | false | Very high quality, near perfect, *recommended*. |
|
44 |
| [Yi-Coder-9B-Chat-Q5_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_L.gguf) | Q5_K_L | 6.42GB | false | Uses Q8_0 for embed and output weights. High quality, *recommended*. |
|
45 |
| [Yi-Coder-9B-Chat-Q5_K_M.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_M.gguf) | Q5_K_M | 6.26GB | false | High quality, *recommended*. |
|
46 |
| [Yi-Coder-9B-Chat-Q5_K_S.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_S.gguf) | Q5_K_S | 6.11GB | false | High quality, *recommended*. |
|
47 |
| [Yi-Coder-9B-Chat-Q4_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_L.gguf) | Q4_K_L | 5.52GB | false | Uses Q8_0 for embed and output weights. Good quality, *recommended*. |
|
48 |
| [Yi-Coder-9B-Chat-Q4_K_M.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_M.gguf) | Q4_K_M | 5.33GB | false | Good quality, default size for must use cases, *recommended*. |
|
49 |
-
| [Yi-Coder-9B-Chat-Q4_K_M.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_M.gguf) | Q4_K_M | 5.33GB | false | Good quality, default size for must use cases, *recommended*. |
|
50 |
| [Yi-Coder-9B-Chat-Q4_K_S.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_S.gguf) | Q4_K_S | 5.07GB | false | Slightly lower quality with more space savings, *recommended*. |
|
51 |
| [Yi-Coder-9B-Chat-Q4_0.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0.gguf) | Q4_0 | 5.05GB | false | Legacy format, generally not worth using over similarly sized formats |
|
52 |
-
| [Yi-Coder-9B-Chat-Q4_0_4_4.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_4_4.gguf) | Q4_0_4_4 | 5.04GB | false | Optimized for ARM inference. Should work well on all ARM chips, pick this if you're unsure. |
|
53 |
| [Yi-Coder-9B-Chat-Q4_0_8_8.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_8_8.gguf) | Q4_0_8_8 | 5.04GB | false | Optimized for ARM inference. Requires 'sve' support (see link below). |
|
54 |
| [Yi-Coder-9B-Chat-Q4_0_4_8.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_4_8.gguf) | Q4_0_4_8 | 5.04GB | false | Optimized for ARM inference. Requires 'i8mm' support (see link below). |
|
55 |
| [Yi-Coder-9B-Chat-Q4_0_4_4.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_4_4.gguf) | Q4_0_4_4 | 5.04GB | false | Optimized for ARM inference. Should work well on all ARM chips, pick this if you're unsure. |
|
|
|
37 |
| -------- | ---------- | --------- | ----- | ----------- |
|
38 |
| [Yi-Coder-9B-Chat-f16.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-f16.gguf) | f16 | 17.66GB | false | Full F16 weights. |
|
39 |
| [Yi-Coder-9B-Chat-Q8_0.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q8_0.gguf) | Q8_0 | 9.38GB | false | Extremely high quality, generally unneeded but max available quant. |
|
|
|
40 |
| [Yi-Coder-9B-Chat-Q6_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q6_K_L.gguf) | Q6_K_L | 7.37GB | false | Uses Q8_0 for embed and output weights. Very high quality, near perfect, *recommended*. |
|
41 |
| [Yi-Coder-9B-Chat-Q6_K.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q6_K.gguf) | Q6_K | 7.25GB | false | Very high quality, near perfect, *recommended*. |
|
|
|
42 |
| [Yi-Coder-9B-Chat-Q5_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_L.gguf) | Q5_K_L | 6.42GB | false | Uses Q8_0 for embed and output weights. High quality, *recommended*. |
|
43 |
| [Yi-Coder-9B-Chat-Q5_K_M.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_M.gguf) | Q5_K_M | 6.26GB | false | High quality, *recommended*. |
|
44 |
| [Yi-Coder-9B-Chat-Q5_K_S.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_S.gguf) | Q5_K_S | 6.11GB | false | High quality, *recommended*. |
|
45 |
| [Yi-Coder-9B-Chat-Q4_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_L.gguf) | Q4_K_L | 5.52GB | false | Uses Q8_0 for embed and output weights. Good quality, *recommended*. |
|
46 |
| [Yi-Coder-9B-Chat-Q4_K_M.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_M.gguf) | Q4_K_M | 5.33GB | false | Good quality, default size for must use cases, *recommended*. |
|
|
|
47 |
| [Yi-Coder-9B-Chat-Q4_K_S.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_S.gguf) | Q4_K_S | 5.07GB | false | Slightly lower quality with more space savings, *recommended*. |
|
48 |
| [Yi-Coder-9B-Chat-Q4_0.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0.gguf) | Q4_0 | 5.05GB | false | Legacy format, generally not worth using over similarly sized formats |
|
|
|
49 |
| [Yi-Coder-9B-Chat-Q4_0_8_8.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_8_8.gguf) | Q4_0_8_8 | 5.04GB | false | Optimized for ARM inference. Requires 'sve' support (see link below). |
|
50 |
| [Yi-Coder-9B-Chat-Q4_0_4_8.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_4_8.gguf) | Q4_0_4_8 | 5.04GB | false | Optimized for ARM inference. Requires 'i8mm' support (see link below). |
|
51 |
| [Yi-Coder-9B-Chat-Q4_0_4_4.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_4_4.gguf) | Q4_0_4_4 | 5.04GB | false | Optimized for ARM inference. Should work well on all ARM chips, pick this if you're unsure. |
|