cjpais commited on
Commit
c30ab86
1 Parent(s): 1228c82

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -4,18 +4,18 @@ license: apache-2.0
4
 
5
  # GGUF Quantized LLaVA 1.6 Mistral 7B
6
 
7
- Notes: Was prepared with a unofficial script, and is likely missing some data and lacking some performance. Will update quants when better script is available
8
 
9
  ## Provided files
10
  | Name | Quant method | Bits | Size | Use case |
11
  | ---- | ---- | ---- | ---- | ----- |
12
  | [llava-v1.6-mistral-7b.Q3_K_XS.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q3_K_XS.gguf) | Q3_K_XS | 3 | 2.99 GB| very small, high quality loss |
13
- | [llava-v1.6-mistral-7b.Q3_K_M.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q3_K.gguf) | Q3_K_M | 3 | 3.52 GB| very small, high quality loss |
14
  | [llava-v1.6-mistral-7b.Q4_K_M.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q4_K_M.gguf) | Q4_K_M | 4 | 4.37 GB| medium, balanced quality - recommended |
15
  | [llava-v1.6-mistral-7b.Q5_K_S.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q5_K_S.gguf) | Q5_K_S | 5 | 5.00 GB| large, low quality loss - recommended |
16
  | [llava-v1.6-mistral-7b.Q5_K_M.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q5_K_M.gguf) | Q5_K_M | 5 | 5.13 GB| large, very low quality loss - recommended |
17
- | [llava-v1.6-mistral-7b.Q6_K.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-1.6-mistral-7b.Q6_K.gguf) | Q6_K | 6 | 5.94 GB| very large, extremely low quality loss |
18
- | [llava-v1.6-mistral-7b.Q8_0.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-1.6-mistral-7b.Q8_0.gguf) | Q8_0 | 8 | 7.7 GB| very large, extremely low quality loss - not recommended |
19
 
20
  <br>
21
  <br>
 
4
 
5
  # GGUF Quantized LLaVA 1.6 Mistral 7B
6
 
7
+ Updated quants and projector from [PR #5267](https://github.com/ggerganov/llama.cpp/pull/5267)
8
 
9
  ## Provided files
10
  | Name | Quant method | Bits | Size | Use case |
11
  | ---- | ---- | ---- | ---- | ----- |
12
  | [llava-v1.6-mistral-7b.Q3_K_XS.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q3_K_XS.gguf) | Q3_K_XS | 3 | 2.99 GB| very small, high quality loss |
13
+ | [llava-v1.6-mistral-7b.Q3_K_M.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q3_K_M.gguf) | Q3_K_M | 3 | 3.52 GB| very small, high quality loss |
14
  | [llava-v1.6-mistral-7b.Q4_K_M.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q4_K_M.gguf) | Q4_K_M | 4 | 4.37 GB| medium, balanced quality - recommended |
15
  | [llava-v1.6-mistral-7b.Q5_K_S.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q5_K_S.gguf) | Q5_K_S | 5 | 5.00 GB| large, low quality loss - recommended |
16
  | [llava-v1.6-mistral-7b.Q5_K_M.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q5_K_M.gguf) | Q5_K_M | 5 | 5.13 GB| large, very low quality loss - recommended |
17
+ | [llava-v1.6-mistral-7b.Q6_K.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q6_K.gguf) | Q6_K | 6 | 5.94 GB| very large, extremely low quality loss |
18
+ | [llava-v1.6-mistral-7b.Q8_0.gguf](https://huggingface.co/cjpais/llava-1.6-mistral-7b-gguf/blob/main/llava-v1.6-mistral-7b.Q8_0.gguf) | Q8_0 | 8 | 7.7 GB| very large, extremely low quality loss - not recommended |
19
 
20
  <br>
21
  <br>