Frederik Gaasdal Jensen • Henry Stoll • Sippo Rossi • Raghava Rao Mukkamala
UNHCR Hate Speech Detection Model
This is a transformer model that can detect hate and offensive speech for English text. The primary use-case of this model is to detect hate speech targeted at refugees. The model is based on roberta-uncased and was fine-tuned on 12 abusive language datasets.
The model has been developed as a collaboration between UNHCR, the UN Refugee Agency, and Copenhagen Business School.
- F1-score on test set (10% of the overall dataset): 81%
- Hatecheck score: 90.3%
Labels
{
0: "Normal",
1: "Offensive",
2: "Hate speech",
}
- Downloads last month
- 1,261
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.