maddes8cht
commited on
Commit
•
04a8a25
1
Parent(s):
094a9ad
"Update README.md"
Browse files
README.md
CHANGED
@@ -17,9 +17,9 @@ I'm constantly enhancing these model descriptions to provide you with the most r
|
|
17 |
|
18 |
# K-Quants in Falcon 7b models
|
19 |
|
20 |
-
New Llama.cpp
|
21 |
|
22 |
-
For Falcon 7B models, although only a quarter of the layers can be quantized with true K-quants, this approach still benefits from utilizing
|
23 |
|
24 |
So this solution ensures improved performance and efficiency over legacy Q4_0, Q4_1, Q5_0 and Q5_1 Quantizations.
|
25 |
|
@@ -28,14 +28,12 @@ So this solution ensures improved performance and efficiency over legacy Q4_0, Q
|
|
28 |
|
29 |
As previously noted on the [Llama.cpp GitHub repository](https://github.com/ggerganov/llama.cpp#hot-topics), all new Llama.cpp releases after October 18, 2023, required re-quantization due to the implementation of the new BPE tokenizer.
|
30 |
|
31 |
-
|
32 |
-
|
33 |
-
**Key Points:**
|
34 |
|
35 |
- **Stay Informed:** Keep an eye on software application release schedules using llama.cpp libraries.
|
36 |
- **Monitor Upload Times:** Re-quantization is complete. Watch for updates on my Hugging Face Model pages.
|
37 |
|
38 |
-
This change
|
39 |
|
40 |
|
41 |
|
|
|
17 |
|
18 |
# K-Quants in Falcon 7b models
|
19 |
|
20 |
+
New releases of Llama.cpp now support K-quantization for previously incompatible models, in particular all Falcon 7B models. This is achieved by employing a fallback solution for model layers that cannot be quantized with real K-quants.
|
21 |
|
22 |
+
For Falcon 7B models, although only a quarter of the layers can be quantized with true K-quants, this approach still benefits from utilizing *different* legacy quantization types Q4_0, Q4_1, Q5_0, and Q5_1. As a result, it offers better quality at the same file size or smaller file sizes with comparable performance.
|
23 |
|
24 |
So this solution ensures improved performance and efficiency over legacy Q4_0, Q4_1, Q5_0 and Q5_1 Quantizations.
|
25 |
|
|
|
28 |
|
29 |
As previously noted on the [Llama.cpp GitHub repository](https://github.com/ggerganov/llama.cpp#hot-topics), all new Llama.cpp releases after October 18, 2023, required re-quantization due to the implementation of the new BPE tokenizer.
|
30 |
|
31 |
+
This re-quantization process for Falcon Models is now complete, the latest quantized models are available here for download. To ensure continued compatibility with recent llama.cpp software, You need to update your Falcon models.
|
|
|
|
|
32 |
|
33 |
- **Stay Informed:** Keep an eye on software application release schedules using llama.cpp libraries.
|
34 |
- **Monitor Upload Times:** Re-quantization is complete. Watch for updates on my Hugging Face Model pages.
|
35 |
|
36 |
+
This change only affects **Falcon** and **Starcoder** models, with other models remaining unaffected.
|
37 |
|
38 |
|
39 |
|