sayhan commited on
Commit
dc1fe92
1 Parent(s): b42d734

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +18 -1
README.md CHANGED
@@ -22,10 +22,27 @@ This repo contains GGUF format model files for [malhajar's Mistral 7B Instruct v
22
  - **Language(s) (NLP):** Turkish
23
  - **Finetuned from model:** [`mistralai/Mistral-7B-Instruct-v0.2`](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
24
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  ## Prompt Template
26
  ```
27
  ### Instruction:
28
  <prompt> (without the <>)
29
  ### Response:
30
  ```
31
- <!-- description end -->
 
22
  - **Language(s) (NLP):** Turkish
23
  - **Finetuned from model:** [`mistralai/Mistral-7B-Instruct-v0.2`](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2)
24
 
25
+ # Quantization methods
26
+ | quantization method | bits | size | use case | recommended |
27
+ |---------------------|------|----------|-----------------------------------------------------|-------------|
28
+ | Q2_K | 2 | 2.72 GB | smallest, significant quality loss | ❌ |
29
+ | Q3_K_S | 3 | 3.16 GB | very small, high quality loss | ❌ |
30
+ | Q3_K_M | 3 | 3.52 GB | very small, high quality loss | ❌ |
31
+ | Q3_K_L | 3 | 3.82 GB | small, substantial quality loss | ❌ |
32
+ | Q4_0 | 4 | 4.11 GB | legacy; small, very high quality loss | ❌ |
33
+ | Q4_K_S | 4 | 4.14 GB | small, greater quality loss | ❌ |
34
+ | Q4_K_M | 4 | 4.37 GB | medium, balanced quality | ✅ |
35
+ | Q5_0 | 5 | 5.00 GB | legacy; medium, balanced quality | ❌ |
36
+ | Q5_K_S | 5 | 5.00 GB | large, low quality loss | ✅ |
37
+ | Q5_K_M | 5 | 5.13 GB | large, very low quality loss | ✅ |
38
+ | Q6_K | 6 | 5.94 GB | very large, extremely low quality loss | ❌ |
39
+ | Q8_0 | 8 | 7.70 GB | very large, extremely low quality loss | ❌ |
40
+ | FP16 | 16 | 14.5 GB | enormous, minuscule quality loss | ❌ |
41
+
42
  ## Prompt Template
43
  ```
44
  ### Instruction:
45
  <prompt> (without the <>)
46
  ### Response:
47
  ```
48
+ <!-- description end -->