|
--- |
|
{} |
|
--- |
|
|
|
# HarmAug: Effective Data Augmentation for Knowledge Distillation of Safety Guard Models |
|
|
|
|
|
 |
|
|
|
 |
|
|
|
|
|
This model is a Guard Model, specifically designed to classify the safety of LLM conversations. |
|
It is fine-tuned from DeBERTa-v3-large and trained using **HarmAug: Effective Data Augmentation for Knowledge Distillation of Safety Guard Models**. |
|
The training process involves knowledge distillation paired with data augmentation, using our [**HarmAug Generated Dataset**](https://huggingface.co/datasets/AnonHB/HarmAug_generated_dataset). |
|
|
|
|
|
For more information, please refer to our [anonymous github](https://anonymous.4open.science/r/HarmAug/) |