GGUF
maddes8cht commited on
Commit
1673a59
1 Parent(s): 3ad6112

"Update README.md"

Browse files
Files changed (1) hide show
  1. README.md +19 -10
README.md CHANGED
@@ -17,20 +17,27 @@ I'm constantly enhancing these model descriptions to provide you with the most r
17
  - Model creator: [tiiuae](https://huggingface.co/tiiuae)
18
  - Original model: [falcon-40b](https://huggingface.co/tiiuae/falcon-40b)
19
 
 
 
 
 
 
 
 
 
 
20
  # Important Update for Falcon Models in llama.cpp Versions After October 18, 2023
21
 
22
- As noted on the [Llama.cpp GitHub repository](https://github.com/ggerganov/llama.cpp#hot-topics), all new Llama.cpp releases after October 18, 2023, will require a re-quantization due to the new BPE tokenizer.
23
 
24
- **Good news!** I am glad that my re-quantization process for Falcon Models is nearly complete. Download the latest quantized models to ensure compatibility with recent llama.cpp software.
25
 
26
  **Key Points:**
27
 
28
  - **Stay Informed:** Keep an eye on software application release schedules using llama.cpp libraries.
29
- - **Monitor Upload Times:** Re-quantization is *almost* done. Watch for updates on my Hugging Face Model pages.
30
-
31
- **Important Compatibility Note:** Old software will work with old Falcon models, but expect updated software to exclusively support the new models.
32
 
33
- This change primarily affects **Falcon** and **Starcoder** models, with other models remaining unaffected.
34
 
35
 
36
 
@@ -53,19 +60,21 @@ The core project making use of the ggml library is the [llama.cpp](https://githu
53
 
54
  # Quantization variants
55
 
56
- There is a bunch of quantized files available. How to choose the best for you:
57
 
58
  # Legacy quants
59
 
60
  Q4_0, Q4_1, Q5_0, Q5_1 and Q8 are `legacy` quantization types.
61
  Nevertheless, they are fully supported, as there are several circumstances that cause certain model not to be compatible with the modern K-quants.
62
- Falcon 7B models cannot be quantized to K-quants.
 
 
63
 
64
  # K-quants
65
 
66
- K-quants are based on the idea that the quantization of certain parts affects the quality in different ways. If you quantize certain parts more and others less, you get a more powerful model with the same file size, or a smaller file size and lower memory load with comparable performance.
67
  So, if possible, use K-quants.
68
- With a Q6_K you should find it really hard to find a quality difference to the original model - ask your model two times the same question and you may encounter bigger quality differences.
69
 
70
 
71
 
 
17
  - Model creator: [tiiuae](https://huggingface.co/tiiuae)
18
  - Original model: [falcon-40b](https://huggingface.co/tiiuae/falcon-40b)
19
 
20
+ # K-Quants in Falcon 7b models
21
+
22
+ New Llama.cpp releases now allow for K-quantization of models that were previously incompatible with K-quants. This is achieved by employing a fallback solution for model layers that cannot be accurately quantized with K-quants.
23
+
24
+ For Falcon 7B models, although only a quarter of the layers can be quantized with true K-quants, this approach still benefits from utilizing various legacy quantization types, such as Q4_0, Q4_1, Q5_0, and Q5_1. As a result, it offers better quality at the same file size or smaller file sizes with comparable performance.
25
+
26
+ So this solution ensures improved performance and efficiency over legacy Q4_0, Q4_1, Q5_0 and Q5_1 Quantizations.
27
+
28
+
29
  # Important Update for Falcon Models in llama.cpp Versions After October 18, 2023
30
 
31
+ As previously noted on the [Llama.cpp GitHub repository](https://github.com/ggerganov/llama.cpp#hot-topics), all new Llama.cpp releases after October 18, 2023, required re-quantization due to the implementation of the new BPE tokenizer.
32
 
33
+ **Update:** The re-quantization process for Falcon Models is now complete, and the latest quantized models are available for download. To ensure continued compatibility with recent llama.cpp software, You need to update your Falcon models.
34
 
35
  **Key Points:**
36
 
37
  - **Stay Informed:** Keep an eye on software application release schedules using llama.cpp libraries.
38
+ - **Monitor Upload Times:** Re-quantization is complete. Watch for updates on my Hugging Face Model pages.
 
 
39
 
40
+ This change primarily affects **Falcon** and **Starcoder** models, with other models remaining unaffected. If you haven't already, please update your Falcon models for seamless compatibility with the latest llama.cpp versions.
41
 
42
 
43
 
 
60
 
61
  # Quantization variants
62
 
63
+ There is a bunch of quantized files available to cater to your specific needs. Here's how to choose the best option for you:
64
 
65
  # Legacy quants
66
 
67
  Q4_0, Q4_1, Q5_0, Q5_1 and Q8 are `legacy` quantization types.
68
  Nevertheless, they are fully supported, as there are several circumstances that cause certain model not to be compatible with the modern K-quants.
69
+ ## Note:
70
+ Now there's a new option to use K-quants even for previously 'incompatible' models, although this involves some fallback solution that makes them not *real* K-quants. More details can be found in affected model descriptions.
71
+ (This mainly refers to Falcon 7b and Starcoder models)
72
 
73
  # K-quants
74
 
75
+ K-quants are designed with the idea that different levels of quantization in specific parts of the model can optimize performance, file size, and memory load.
76
  So, if possible, use K-quants.
77
+ With a Q6_K, you'll likely find it challenging to discern a quality difference from the original model - ask your model two times the same question and you may encounter bigger quality differences.
78
 
79
 
80