## Model Details Original LlamaGuard 7b model can be found [here](https://huggingface.co/meta-llama/LlamaGuard-7b) Llama-Guard is a 7B parameter [Llama 2](https://arxiv.org/abs/2307.09288)-based input-output safeguard model. It can be used for classifying content in both LLM inputs (prompt classification) and in LLM responses (response classification). It acts as an LLM: it generates text in its output that indicates whether a given prompt or response is safe/unsafe, and if unsafe based on a policy, it also lists the violating subcategories. Here is an example: ![](Llama-Guard_example.png) These are exl2 4.65bpw quantized weights. Original 7B model performs on binary classification of 2k toxic chat test examples Precision: 0.9, Recall: 0.277, F1 Score: 0.424 4.0bpw performs Precision: 0.92, Recall: 0.246, F1 Score: 0.389 4.65bpw performs Precision: 0.903, Recall: 0.256, F1 Score: 0.400