naman632 commited on
Commit
ba1e54f
1 Parent(s): a058972

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -56
README.md CHANGED
@@ -4,60 +4,5 @@ language:
4
  tags:
5
  - text-generation
6
  - conditional-text-generation
 
7
  ---
8
- # Model Details
9
-
10
-
11
- This is a conditional language model based on [gpt2-medium](https://huggingface.co/gpt2-medium/) but with a vocabulary from [t5-base](https://huggingface.co/t5-base), for compatibility with T5-based paraphrasers such as [t5-paranmt-detox](https://huggingface.co/SkolkovoInstitute/t5-paranmt-detox). The model is conditional on two styles, `toxic` and `normal`, and was fine-tuned on the dataset from the Jigsaw [toxic comment classification challenge](https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge).
12
-
13
- The model was trained for the paper [Text Detoxification using Large Pre-trained Neural Models](https://arxiv.org/abs/2109.08914) (Dale et al, 2021) that describes its possible usage in more detail.
14
-
15
- An example of its use and the code for its training is given in https://github.com/skoltech-nlp/detox.
16
-
17
-
18
- ## Model Description
19
-
20
- - **Developed by:** SkolkovoInstitute
21
- - **Model type:** Conditional Text Generation
22
- - **Language:** English
23
- - **Related Models:**
24
- - **Parent Model:** [gpt2-medium](https://huggingface.co/gpt2-medium/)
25
- - **Source of vocabulary:** [t5-base](https://huggingface.co/t5-base)
26
- - **Resources for more information:**
27
- - The paper [Text Detoxification using Large Pre-trained Neural Models](https://arxiv.org/abs/2109.08914)
28
- - Its repository https://github.com/skoltech-nlp/detox.
29
-
30
- # Uses
31
-
32
- The model is intended for usage as a discriminator in a text detoxification pipeline using the ParaGeDi approach (see [the paper](https://arxiv.org/abs/2109.08914) for more details). It can also be used for text generation conditional on toxic or non-toxic style, but we do not know how to condition it on the things other than toxicity, so we do not recommend this usage. Another possible use is as a toxicity classifier (using the Bayes rule), but the model is not expected to perform better than e.g. a BERT-based standard classifier.
33
-
34
- # Bias, Risks, and Limitations
35
- The model inherits all the risks of its parent model, [gpt2-medium](https://huggingface.co/gpt2-medium/). It also inherits all the biases of the [Jigsaw dataset](https://www.kaggle.com/c/jigsaw-toxic-comment-classification-challenge) on which it was fine-tuned. The model is intended to be conditional on style, but in fact it does not clearly separate the concepts of style and content, so it might regard some texts as toxic or safe based not on the style, but on their topics or keywords.
36
-
37
- # Training Details
38
- See the paper [Text Detoxification using Large Pre-trained Neural Models](https://arxiv.org/abs/2109.08914) and [the associated code](https://github.com/s-nlp/detox/tree/main/emnlp2021/style_transfer/paraGeDi).
39
-
40
- # Evaluation
41
- The model has not been evaluated on its own, only as a part as a ParaGeDi text detoxification pipeline (see [the paper](https://arxiv.org/abs/2109.08914)).
42
-
43
- # Citation
44
- **BibTeX:**
45
- ```
46
- @inproceedings{dale-etal-2021-text,
47
- title = "Text Detoxification using Large Pre-trained Neural Models",
48
- author = "Dale, David and
49
- Voronov, Anton and
50
- Dementieva, Daryna and
51
- Logacheva, Varvara and
52
- Kozlova, Olga and
53
- Semenov, Nikita and
54
- Panchenko, Alexander",
55
- booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing",
56
- month = nov,
57
- year = "2021",
58
- address = "Online and Punta Cana, Dominican Republic",
59
- publisher = "Association for Computational Linguistics",
60
- url = "https://aclanthology.org/2021.emnlp-main.629",
61
- pages = "7979--7996",
62
- }
63
- ```
 
4
  tags:
5
  - text-generation
6
  - conditional-text-generation
7
+ - developed by NLP_team for Advanced NLP Project.
8
  ---