pinyuchen commited on
Commit
4ff1f23
1 Parent(s): 21dfe90

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -10
README.md CHANGED
@@ -8,13 +8,14 @@ pipeline_tag: text-classification
8
 
9
  ## Model Details
10
 
11
- RADAR-Vicuna-7B is an AI-text detector trained via adversrial learning between the detector and a paraphraser on human-text corpus([OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext)) and ai-text corpus generated
12
  based on [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext).
13
 
14
  - **Developed by:** [TrustSafeAI](https://huggingface.co/TrustSafeAI)
15
- - **Model type:** An encoder-only language model based on the transformer architecture.
16
- - **License:** Non-commercial license
17
- - **Trained from model:** [RoBERTa](https://arxiv.org/abs/1907.11692).
 
18
 
19
  ### Model Sources
20
 
@@ -24,7 +25,7 @@ based on [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext).
24
 
25
  ## Uses
26
  Users could use this detector to assist them in detecting text generated by large language models.
27
- Please be noted that this detector is trained on ai-text generated by Vicuna-7B-v1.1. As the model only supports [non-commercial use](https://huggingface.co/lmsys/vicuna-7b-v1.1#model-details), the intended users are **not allowed to involve this detector into commercial activities**.
28
 
29
  ## Get Started with the Model
30
  Please refer to the following guidelines to see how to locally run the downloaded model or use our API service hosted on Huggingface Space.
@@ -34,15 +35,15 @@ Please refer to the following guidelines to see how to locally run the downloade
34
  ## Training Pipeline
35
 
36
  We propose adversarial learning between a paraphraser and our detector. The paraphraser's goal is to make the AI-generated text more like human-writen and the detector's goal is to
37
- promote it's ability to identify the ai-text.
38
 
39
- - **(Step 1) Training Data preparation**: Before training, we use Vicuna-7B to generate AI-text by performing text-completion based on the prefix span of human-text in [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext).
40
 
41
- - **(Step 2) Update the paraphraser** During training, the paraphraser will do paraphrasing on the AI-text generated in **Step 1**. And then collect the reward returned by the detector to update the parphraser using Proxy Proximal Optimization loss.
42
 
43
- - **(Step 3) Update the detector** The detector is optimized using the logistic loss on the human-text, ai-text and paraphrased ai-text.
44
 
45
- See more details in the Section 3 and 4 of this [paper](https://arxiv.org/pdf/2307.03838.pdf).
46
 
47
  ## Ethical Considerations
48
  We suggest users use our tool to assist with identifying AI-written content at scale and with discretion. If the detection result is to be used as evidence, further validation steps
 
8
 
9
  ## Model Details
10
 
11
+ RADAR-Vicuna-7B is an AI-text detector trained via adversarial learning between the detector and a paraphraser on human-text corpus ([OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext)) and AI-text corpus generated
12
  based on [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext).
13
 
14
  - **Developed by:** [TrustSafeAI](https://huggingface.co/TrustSafeAI)
15
+ - **Model type:** An encoder-only language model based on the transformer architecture (RoBERTa).
16
+ - **License:** [Non-commercial license](https://huggingface.co/lmsys/vicuna-7b-v1.1#model-details) (inherited from Vicuna-7B-v1.1)
17
+ - **Trained from model:** [RoBERTa](https://arxiv.org/abs/1907.11692)
18
+
19
 
20
  ### Model Sources
21
 
 
25
 
26
  ## Uses
27
  Users could use this detector to assist them in detecting text generated by large language models.
28
+ Please note that this detector is trained on AI-text generated by Vicuna-7B-v1.1. As the model only supports [non-commercial use](https://huggingface.co/lmsys/vicuna-7b-v1.1#model-details), the intended users are **not allowed to involve this detector into commercial activities**.
29
 
30
  ## Get Started with the Model
31
  Please refer to the following guidelines to see how to locally run the downloaded model or use our API service hosted on Huggingface Space.
 
35
  ## Training Pipeline
36
 
37
  We propose adversarial learning between a paraphraser and our detector. The paraphraser's goal is to make the AI-generated text more like human-writen and the detector's goal is to
38
+ promote it's ability to identify the AI-text.
39
 
40
+ - **(Step 1) Training Data preparation**: Before training, we use Vicuna-7B to generate AI-text by performing text completion based on the prefix span of human-text in [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext).
41
 
42
+ - **(Step 2) Update the paraphraser** During training, the paraphraser will do paraphrasing on the AI-text generated in **Step 1**. And then collect the reward returned by the detector to update the paraphraser using Proxy Proximal Optimization loss.
43
 
44
+ - **(Step 3) Update the detector** The detector is optimized using the logistic loss on the human-text, AI-text and paraphrased AI-text.
45
 
46
+ See more details in Sections 3 and 4 of this [paper](https://arxiv.org/pdf/2307.03838.pdf).
47
 
48
  ## Ethical Considerations
49
  We suggest users use our tool to assist with identifying AI-written content at scale and with discretion. If the detection result is to be used as evidence, further validation steps