Update README.md
Browse files
README.md
CHANGED
@@ -18,6 +18,6 @@ language:
|
|
18 |
|
19 |
This is [mistralai/Mistral-Large-Instruct-2407](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407), converted to GGUF and quantized to q8_0. Both the model and the embedding/output tensors are q8_0.
|
20 |
|
21 |
-
The model is split using the llama.cpp/llama-gguf-split cli utility into shards no larger than 7GB. The purpose of this is to make it less painful to resume downloading if interrupted.
|
22 |
|
23 |
[GGUFv3](https://github.com/ggerganov/ggml/blob/master/docs/gguf.md)
|
|
|
18 |
|
19 |
This is [mistralai/Mistral-Large-Instruct-2407](https://huggingface.co/mistralai/Mistral-Large-Instruct-2407), converted to GGUF and quantized to q8_0. Both the model and the embedding/output tensors are q8_0.
|
20 |
|
21 |
+
The model is split using the `llama.cpp/llama-gguf-split` cli utility into shards no larger than 7GB. The purpose of this is to make it less painful to resume downloading if interrupted.
|
22 |
|
23 |
[GGUFv3](https://github.com/ggerganov/ggml/blob/master/docs/gguf.md)
|