Update README.md
Browse files
README.md
CHANGED
@@ -26,6 +26,8 @@ base_model:
|
|
26 |
This is a quantized model from the llama guard series for use with [exllamav2](https://github.com/turboderp/exllamav2).\
|
27 |
Please refer to the original model card for instructions on how to use this model.
|
28 |
|
|
|
|
|
29 |
### BPW:
|
30 |
|
31 |
[4.0](https://huggingface.co/Anthonyg5005/Llama-Guard-3-1B-exl2/tree/4.0bpw)\
|
|
|
26 |
This is a quantized model from the llama guard series for use with [exllamav2](https://github.com/turboderp/exllamav2).\
|
27 |
Please refer to the original model card for instructions on how to use this model.
|
28 |
|
29 |
+
### Only use this version for testing, BF16 weights recommended for deployment or at least 8bpw.
|
30 |
+
|
31 |
### BPW:
|
32 |
|
33 |
[4.0](https://huggingface.co/Anthonyg5005/Llama-Guard-3-1B-exl2/tree/4.0bpw)\
|