maddes8cht
commited on
Commit
•
1673a59
1
Parent(s):
3ad6112
"Update README.md"
Browse files
README.md
CHANGED
@@ -17,20 +17,27 @@ I'm constantly enhancing these model descriptions to provide you with the most r
|
|
17 |
- Model creator: [tiiuae](https://huggingface.co/tiiuae)
|
18 |
- Original model: [falcon-40b](https://huggingface.co/tiiuae/falcon-40b)
|
19 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
# Important Update for Falcon Models in llama.cpp Versions After October 18, 2023
|
21 |
|
22 |
-
As noted on the [Llama.cpp GitHub repository](https://github.com/ggerganov/llama.cpp#hot-topics), all new Llama.cpp releases after October 18, 2023,
|
23 |
|
24 |
-
**
|
25 |
|
26 |
**Key Points:**
|
27 |
|
28 |
- **Stay Informed:** Keep an eye on software application release schedules using llama.cpp libraries.
|
29 |
-
- **Monitor Upload Times:** Re-quantization is
|
30 |
-
|
31 |
-
**Important Compatibility Note:** Old software will work with old Falcon models, but expect updated software to exclusively support the new models.
|
32 |
|
33 |
-
This change primarily affects **Falcon** and **Starcoder** models, with other models remaining unaffected.
|
34 |
|
35 |
|
36 |
|
@@ -53,19 +60,21 @@ The core project making use of the ggml library is the [llama.cpp](https://githu
|
|
53 |
|
54 |
# Quantization variants
|
55 |
|
56 |
-
There is a bunch of quantized files available.
|
57 |
|
58 |
# Legacy quants
|
59 |
|
60 |
Q4_0, Q4_1, Q5_0, Q5_1 and Q8 are `legacy` quantization types.
|
61 |
Nevertheless, they are fully supported, as there are several circumstances that cause certain model not to be compatible with the modern K-quants.
|
62 |
-
|
|
|
|
|
63 |
|
64 |
# K-quants
|
65 |
|
66 |
-
K-quants are
|
67 |
So, if possible, use K-quants.
|
68 |
-
With a Q6_K you
|
69 |
|
70 |
|
71 |
|
|
|
17 |
- Model creator: [tiiuae](https://huggingface.co/tiiuae)
|
18 |
- Original model: [falcon-40b](https://huggingface.co/tiiuae/falcon-40b)
|
19 |
|
20 |
+
# K-Quants in Falcon 7b models
|
21 |
+
|
22 |
+
New Llama.cpp releases now allow for K-quantization of models that were previously incompatible with K-quants. This is achieved by employing a fallback solution for model layers that cannot be accurately quantized with K-quants.
|
23 |
+
|
24 |
+
For Falcon 7B models, although only a quarter of the layers can be quantized with true K-quants, this approach still benefits from utilizing various legacy quantization types, such as Q4_0, Q4_1, Q5_0, and Q5_1. As a result, it offers better quality at the same file size or smaller file sizes with comparable performance.
|
25 |
+
|
26 |
+
So this solution ensures improved performance and efficiency over legacy Q4_0, Q4_1, Q5_0 and Q5_1 Quantizations.
|
27 |
+
|
28 |
+
|
29 |
# Important Update for Falcon Models in llama.cpp Versions After October 18, 2023
|
30 |
|
31 |
+
As previously noted on the [Llama.cpp GitHub repository](https://github.com/ggerganov/llama.cpp#hot-topics), all new Llama.cpp releases after October 18, 2023, required re-quantization due to the implementation of the new BPE tokenizer.
|
32 |
|
33 |
+
**Update:** The re-quantization process for Falcon Models is now complete, and the latest quantized models are available for download. To ensure continued compatibility with recent llama.cpp software, You need to update your Falcon models.
|
34 |
|
35 |
**Key Points:**
|
36 |
|
37 |
- **Stay Informed:** Keep an eye on software application release schedules using llama.cpp libraries.
|
38 |
+
- **Monitor Upload Times:** Re-quantization is complete. Watch for updates on my Hugging Face Model pages.
|
|
|
|
|
39 |
|
40 |
+
This change primarily affects **Falcon** and **Starcoder** models, with other models remaining unaffected. If you haven't already, please update your Falcon models for seamless compatibility with the latest llama.cpp versions.
|
41 |
|
42 |
|
43 |
|
|
|
60 |
|
61 |
# Quantization variants
|
62 |
|
63 |
+
There is a bunch of quantized files available to cater to your specific needs. Here's how to choose the best option for you:
|
64 |
|
65 |
# Legacy quants
|
66 |
|
67 |
Q4_0, Q4_1, Q5_0, Q5_1 and Q8 are `legacy` quantization types.
|
68 |
Nevertheless, they are fully supported, as there are several circumstances that cause certain model not to be compatible with the modern K-quants.
|
69 |
+
## Note:
|
70 |
+
Now there's a new option to use K-quants even for previously 'incompatible' models, although this involves some fallback solution that makes them not *real* K-quants. More details can be found in affected model descriptions.
|
71 |
+
(This mainly refers to Falcon 7b and Starcoder models)
|
72 |
|
73 |
# K-quants
|
74 |
|
75 |
+
K-quants are designed with the idea that different levels of quantization in specific parts of the model can optimize performance, file size, and memory load.
|
76 |
So, if possible, use K-quants.
|
77 |
+
With a Q6_K, you'll likely find it challenging to discern a quality difference from the original model - ask your model two times the same question and you may encounter bigger quality differences.
|
78 |
|
79 |
|
80 |
|