naman632 commited on
Commit
ffd1c9a
1 Parent(s): ecb9f85

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -31
README.md CHANGED
@@ -1,36 +1,7 @@
1
- ---
2
- language:
3
- - en
4
- tags:
5
- - toxic comments classification
6
- licenses:
7
- - cc-by-nc-sa
8
- ---
9
 
10
  ## Toxicity Classification Model
11
 
12
  This model is trained for toxicity classification task. The dataset used for training is the merge of the English parts of the three datasets by **Jigsaw** ([Jigsaw 2018](https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge), [Jigsaw 2019](https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification), [Jigsaw 2020](https://www.kaggle.com/c/jigsaw-multilingual-toxic-comment-classification)), containing around 2 million examples. We split it into two parts and fine-tune a RoBERTa model ([RoBERTa: A Robustly Optimized BERT Pretraining Approach](https://arxiv.org/abs/1907.11692)) on it. The classifiers perform closely on the test set of the first Jigsaw competition, reaching the **AUC-ROC** of 0.98 and **F1-score** of 0.76.
13
 
14
- ## How to use
15
- ```python
16
- from transformers import RobertaTokenizer, RobertaForSequenceClassification
17
-
18
- # load tokenizer and model weights
19
- tokenizer = RobertaTokenizer.from_pretrained('SkolkovoInstitute/roberta_toxicity_classifier')
20
- model = RobertaForSequenceClassification.from_pretrained('SkolkovoInstitute/roberta_toxicity_classifier')
21
-
22
- # prepare the input
23
- batch = tokenizer.encode('you are amazing', return_tensors='pt')
24
-
25
- # inference
26
- model(batch)
27
- ```
28
-
29
- ## Licensing Information
30
-
31
- [Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License][cc-by-nc-sa].
32
-
33
- [![CC BY-NC-SA 4.0][cc-by-nc-sa-image]][cc-by-nc-sa]
34
-
35
- [cc-by-nc-sa]: http://creativecommons.org/licenses/by-nc-sa/4.0/
36
- [cc-by-nc-sa-image]: https://i.creativecommons.org/l/by-nc-sa/4.0/88x31.png
 
1
+ This model is trained by NLP_team for the Advanced NLP course, 2022.
2
+ The model was trained for the paper [Text Detoxification using Large Pre-trained Neural Models](https://arxiv.org/abs/1911.00536).
 
 
 
 
 
 
3
 
4
  ## Toxicity Classification Model
5
 
6
  This model is trained for toxicity classification task. The dataset used for training is the merge of the English parts of the three datasets by **Jigsaw** ([Jigsaw 2018](https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge), [Jigsaw 2019](https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification), [Jigsaw 2020](https://www.kaggle.com/c/jigsaw-multilingual-toxic-comment-classification)), containing around 2 million examples. We split it into two parts and fine-tune a RoBERTa model ([RoBERTa: A Robustly Optimized BERT Pretraining Approach](https://arxiv.org/abs/1907.11692)) on it. The classifiers perform closely on the test set of the first Jigsaw competition, reaching the **AUC-ROC** of 0.98 and **F1-score** of 0.76.
7