bartowski commited on
Commit
2f28bfc
1 Parent(s): 673634f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -4
README.md CHANGED
@@ -37,19 +37,15 @@ Fixing tokenizer
37
  | -------- | ---------- | --------- | ----- | ----------- |
38
  | [Yi-Coder-9B-Chat-f16.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-f16.gguf) | f16 | 17.66GB | false | Full F16 weights. |
39
  | [Yi-Coder-9B-Chat-Q8_0.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q8_0.gguf) | Q8_0 | 9.38GB | false | Extremely high quality, generally unneeded but max available quant. |
40
- | [Yi-Coder-9B-Chat-Q8_0.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q8_0.gguf) | Q8_0 | 9.38GB | false | Extremely high quality, generally unneeded but max available quant. |
41
  | [Yi-Coder-9B-Chat-Q6_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q6_K_L.gguf) | Q6_K_L | 7.37GB | false | Uses Q8_0 for embed and output weights. Very high quality, near perfect, *recommended*. |
42
  | [Yi-Coder-9B-Chat-Q6_K.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q6_K.gguf) | Q6_K | 7.25GB | false | Very high quality, near perfect, *recommended*. |
43
- | [Yi-Coder-9B-Chat-Q6_K.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q6_K.gguf) | Q6_K | 7.25GB | false | Very high quality, near perfect, *recommended*. |
44
  | [Yi-Coder-9B-Chat-Q5_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_L.gguf) | Q5_K_L | 6.42GB | false | Uses Q8_0 for embed and output weights. High quality, *recommended*. |
45
  | [Yi-Coder-9B-Chat-Q5_K_M.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_M.gguf) | Q5_K_M | 6.26GB | false | High quality, *recommended*. |
46
  | [Yi-Coder-9B-Chat-Q5_K_S.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_S.gguf) | Q5_K_S | 6.11GB | false | High quality, *recommended*. |
47
  | [Yi-Coder-9B-Chat-Q4_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_L.gguf) | Q4_K_L | 5.52GB | false | Uses Q8_0 for embed and output weights. Good quality, *recommended*. |
48
  | [Yi-Coder-9B-Chat-Q4_K_M.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_M.gguf) | Q4_K_M | 5.33GB | false | Good quality, default size for must use cases, *recommended*. |
49
- | [Yi-Coder-9B-Chat-Q4_K_M.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_M.gguf) | Q4_K_M | 5.33GB | false | Good quality, default size for must use cases, *recommended*. |
50
  | [Yi-Coder-9B-Chat-Q4_K_S.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_S.gguf) | Q4_K_S | 5.07GB | false | Slightly lower quality with more space savings, *recommended*. |
51
  | [Yi-Coder-9B-Chat-Q4_0.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0.gguf) | Q4_0 | 5.05GB | false | Legacy format, generally not worth using over similarly sized formats |
52
- | [Yi-Coder-9B-Chat-Q4_0_4_4.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_4_4.gguf) | Q4_0_4_4 | 5.04GB | false | Optimized for ARM inference. Should work well on all ARM chips, pick this if you're unsure. |
53
  | [Yi-Coder-9B-Chat-Q4_0_8_8.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_8_8.gguf) | Q4_0_8_8 | 5.04GB | false | Optimized for ARM inference. Requires 'sve' support (see link below). |
54
  | [Yi-Coder-9B-Chat-Q4_0_4_8.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_4_8.gguf) | Q4_0_4_8 | 5.04GB | false | Optimized for ARM inference. Requires 'i8mm' support (see link below). |
55
  | [Yi-Coder-9B-Chat-Q4_0_4_4.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_4_4.gguf) | Q4_0_4_4 | 5.04GB | false | Optimized for ARM inference. Should work well on all ARM chips, pick this if you're unsure. |
 
37
  | -------- | ---------- | --------- | ----- | ----------- |
38
  | [Yi-Coder-9B-Chat-f16.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-f16.gguf) | f16 | 17.66GB | false | Full F16 weights. |
39
  | [Yi-Coder-9B-Chat-Q8_0.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q8_0.gguf) | Q8_0 | 9.38GB | false | Extremely high quality, generally unneeded but max available quant. |
 
40
  | [Yi-Coder-9B-Chat-Q6_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q6_K_L.gguf) | Q6_K_L | 7.37GB | false | Uses Q8_0 for embed and output weights. Very high quality, near perfect, *recommended*. |
41
  | [Yi-Coder-9B-Chat-Q6_K.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q6_K.gguf) | Q6_K | 7.25GB | false | Very high quality, near perfect, *recommended*. |
 
42
  | [Yi-Coder-9B-Chat-Q5_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_L.gguf) | Q5_K_L | 6.42GB | false | Uses Q8_0 for embed and output weights. High quality, *recommended*. |
43
  | [Yi-Coder-9B-Chat-Q5_K_M.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_M.gguf) | Q5_K_M | 6.26GB | false | High quality, *recommended*. |
44
  | [Yi-Coder-9B-Chat-Q5_K_S.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q5_K_S.gguf) | Q5_K_S | 6.11GB | false | High quality, *recommended*. |
45
  | [Yi-Coder-9B-Chat-Q4_K_L.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_L.gguf) | Q4_K_L | 5.52GB | false | Uses Q8_0 for embed and output weights. Good quality, *recommended*. |
46
  | [Yi-Coder-9B-Chat-Q4_K_M.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_M.gguf) | Q4_K_M | 5.33GB | false | Good quality, default size for must use cases, *recommended*. |
 
47
  | [Yi-Coder-9B-Chat-Q4_K_S.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_K_S.gguf) | Q4_K_S | 5.07GB | false | Slightly lower quality with more space savings, *recommended*. |
48
  | [Yi-Coder-9B-Chat-Q4_0.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0.gguf) | Q4_0 | 5.05GB | false | Legacy format, generally not worth using over similarly sized formats |
 
49
  | [Yi-Coder-9B-Chat-Q4_0_8_8.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_8_8.gguf) | Q4_0_8_8 | 5.04GB | false | Optimized for ARM inference. Requires 'sve' support (see link below). |
50
  | [Yi-Coder-9B-Chat-Q4_0_4_8.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_4_8.gguf) | Q4_0_4_8 | 5.04GB | false | Optimized for ARM inference. Requires 'i8mm' support (see link below). |
51
  | [Yi-Coder-9B-Chat-Q4_0_4_4.gguf](https://huggingface.co/bartowski/Yi-Coder-9B-Chat-GGUF/blob/main/Yi-Coder-9B-Chat-Q4_0_4_4.gguf) | Q4_0_4_4 | 5.04GB | false | Optimized for ARM inference. Should work well on all ARM chips, pick this if you're unsure. |