BeaverTails Classifiers
Collection
Safety classifiers fine-tuned on a bilingual dataset composed of the English QA pairs from BeaverTails and the Italian QA pairs from BeaverTails-IT.
•
3 items
•
Updated
This model is a fine-tuned version of meta-llama/Llama-3.1-8B-Instruct on the saiteki-kai/BeaverTails-it dataset. It achieves the following results on the evaluation set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
Training Loss | Epoch | Step | Validation Loss | Accuracy | Macro F1 | Macro Precision | Macro Recall | Micro F1 | Micro Precision | Micro Recall | Flagged/accuracy | Flagged/precision | Flagged/recall | Flagged/f1 |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0.0688 | 1.0 | 8454 | 0.0799 | 0.6891 | 0.6367 | 0.7276 | 0.5931 | 0.7464 | 0.8015 | 0.6984 | 0.8491 | 0.8948 | 0.8260 | 0.8590 |
0.0745 | 2.0 | 16908 | 0.0777 | 0.6956 | 0.6295 | 0.7647 | 0.5680 | 0.7503 | 0.8171 | 0.6935 | 0.8532 | 0.9108 | 0.8160 | 0.8608 |
0.06 | 3.0 | 25362 | 0.0781 | 0.6965 | 0.6444 | 0.7361 | 0.5968 | 0.7539 | 0.8035 | 0.7100 | 0.8561 | 0.9050 | 0.8284 | 0.8650 |
Base model
meta-llama/Llama-3.1-8B