Update README.md
Browse files
README.md
CHANGED
@@ -1,36 +1,7 @@
|
|
1 |
-
|
2 |
-
|
3 |
-
- en
|
4 |
-
tags:
|
5 |
-
- toxic comments classification
|
6 |
-
licenses:
|
7 |
-
- cc-by-nc-sa
|
8 |
-
---
|
9 |
|
10 |
## Toxicity Classification Model
|
11 |
|
12 |
This model is trained for toxicity classification task. The dataset used for training is the merge of the English parts of the three datasets by **Jigsaw** ([Jigsaw 2018](https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge), [Jigsaw 2019](https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification), [Jigsaw 2020](https://www.kaggle.com/c/jigsaw-multilingual-toxic-comment-classification)), containing around 2 million examples. We split it into two parts and fine-tune a RoBERTa model ([RoBERTa: A Robustly Optimized BERT Pretraining Approach](https://arxiv.org/abs/1907.11692)) on it. The classifiers perform closely on the test set of the first Jigsaw competition, reaching the **AUC-ROC** of 0.98 and **F1-score** of 0.76.
|
13 |
|
14 |
-
## How to use
|
15 |
-
```python
|
16 |
-
from transformers import RobertaTokenizer, RobertaForSequenceClassification
|
17 |
-
|
18 |
-
# load tokenizer and model weights
|
19 |
-
tokenizer = RobertaTokenizer.from_pretrained('SkolkovoInstitute/roberta_toxicity_classifier')
|
20 |
-
model = RobertaForSequenceClassification.from_pretrained('SkolkovoInstitute/roberta_toxicity_classifier')
|
21 |
-
|
22 |
-
# prepare the input
|
23 |
-
batch = tokenizer.encode('you are amazing', return_tensors='pt')
|
24 |
-
|
25 |
-
# inference
|
26 |
-
model(batch)
|
27 |
-
```
|
28 |
-
|
29 |
-
## Licensing Information
|
30 |
-
|
31 |
-
[Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License][cc-by-nc-sa].
|
32 |
-
|
33 |
-
[![CC BY-NC-SA 4.0][cc-by-nc-sa-image]][cc-by-nc-sa]
|
34 |
-
|
35 |
-
[cc-by-nc-sa]: http://creativecommons.org/licenses/by-nc-sa/4.0/
|
36 |
-
[cc-by-nc-sa-image]: https://i.creativecommons.org/l/by-nc-sa/4.0/88x31.png
|
|
|
1 |
+
This model is trained by NLP_team for the Advanced NLP course, 2022.
|
2 |
+
The model was trained for the paper [Text Detoxification using Large Pre-trained Neural Models](https://arxiv.org/abs/1911.00536).
|
|
|
|
|
|
|
|
|
|
|
|
|
3 |
|
4 |
## Toxicity Classification Model
|
5 |
|
6 |
This model is trained for toxicity classification task. The dataset used for training is the merge of the English parts of the three datasets by **Jigsaw** ([Jigsaw 2018](https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge), [Jigsaw 2019](https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification), [Jigsaw 2020](https://www.kaggle.com/c/jigsaw-multilingual-toxic-comment-classification)), containing around 2 million examples. We split it into two parts and fine-tune a RoBERTa model ([RoBERTa: A Robustly Optimized BERT Pretraining Approach](https://arxiv.org/abs/1907.11692)) on it. The classifiers perform closely on the test set of the first Jigsaw competition, reaching the **AUC-ROC** of 0.98 and **F1-score** of 0.76.
|
7 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|