Edward Kim
upload of the 4.65 bpw weights
ecfa6a0
## Model Details
Original LlamaGuard 7b model can be found [here](https://huggingface.co/meta-llama/LlamaGuard-7b)
Llama-Guard is a 7B parameter [Llama 2](https://arxiv.org/abs/2307.09288)-based input-output
safeguard model. It can be used for classifying content in both LLM inputs (prompt
classification) and in LLM responses (response classification).
It acts as an LLM: it generates text in its output that indicates whether a given prompt or
response is safe/unsafe, and if unsafe based on a policy, it also lists the violating subcategories.
Here is an example:
![](Llama-Guard_example.png)
These are exl2 4.65bpw quantized weights.
Original 7B model performs on binary classification of 2k toxic chat test examples
Precision: 0.9, Recall: 0.277, F1 Score: 0.424
4.0bpw performs
Precision: 0.92, Recall: 0.246, F1 Score: 0.389
4.65bpw performs
Precision: 0.903, Recall: 0.256, F1 Score: 0.400