Joseph717171
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -9,7 +9,9 @@ author: Joseph717171 & froggeric (https://huggingface.co/datasets/froggeric/imat
|
|
9 |
|
10 |
# Note: Imatrices uploaded to this repo follow the following naming convention: model-name_training-dataset.imatrix (hyphens are purely used in this example to enhance readability...)
|
11 |
|
12 |
-
# Just download the imatrix for your chosen LLM (Large Language Model), and quantize to your preferred QuantType. (Note the following example already assumes you converted your model to GGUF
|
|
|
|
|
13 |
```
|
14 |
llama.cpp % ./quantize --imatrix path_to_imatrix path_to_model/ggml-model-f16.gguf model_name-QuantType.gguf QuantType
|
15 |
```
|
|
|
9 |
|
10 |
# Note: Imatrices uploaded to this repo follow the following naming convention: model-name_training-dataset.imatrix (hyphens are purely used in this example to enhance readability...)
|
11 |
|
12 |
+
# Just download the imatrix for your chosen LLM (Large Language Model), and quantize to your preferred QuantType. (Note the following example already assumes you converted your model to GGUF)
|
13 |
+
|
14 |
+
# Note: If you need detailed steps to convert your Large Language Model to GGUF, please scroll to the bottom of the page and check out the section: How to convert (Supported) LLMs (Large Language Model) to GGUF format
|
15 |
```
|
16 |
llama.cpp % ./quantize --imatrix path_to_imatrix path_to_model/ggml-model-f16.gguf model_name-QuantType.gguf QuantType
|
17 |
```
|