File size: 1,402 Bytes
ac0454e
 
 
 
 
 
 
 
 
 
7eda77a
ac0454e
76233a5
ac0454e
ff09436
07116f4
ac0454e
 
 
 
 
 
07116f4
ac0454e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
---
datasets:
- Hello-SimpleAI/HC3-Chinese
language:
- zh
pipeline_tag: text-classification
tags:
- chatgpt
---

# Model Card for `Hello-SimpleAI/chatgpt-qa-detector-roberta-chinese`

This model is trained on `question-answer` pairs of **the filtered full-text** from [Hello-SimpleAI/HC3-Chinese](https://huggingface.co/datasets/Hello-SimpleAI/HC3-Chinese).

More details refer to [arxiv: 2301.07597](https://arxiv.org/abs/2301.07597) and Gtihub project [Hello-SimpleAI/chatgpt-comparison-detection](https://github.com/Hello-SimpleAI/chatgpt-comparison-detection).


The base checkpoint is [hfl/chinese-roberta-wwm-ext](https://huggingface.co/hfl/chinese-roberta-wwm-ext).
We train it with all [Hello-SimpleAI/HC3-Chinese](https://huggingface.co/datasets/Hello-SimpleAI/HC3-Chinese) data (without held-out) for 2 epochs.

( 2-epoch is consistent with the experiments in [our paper](https://arxiv.org/abs/2301.07597).)


## Citation

Checkout this papaer [arxiv: 2301.07597](https://arxiv.org/abs/2301.07597)

```
@article{guo-etal-2023-hc3,
    title = "How Close is ChatGPT to Human Experts? Comparison Corpus, Evaluation, and Detection",
    author = "Guo, Biyang  and
      Zhang, Xin  and
      Wang, Ziyuan  and
      Jiang, Minqi  and
      Nie, Jinran  and
      Ding, Yuxuan  and
      Yue, Jianwei  and
      Wu, Yupeng",
    journal={arXiv preprint arxiv:2301.07597}
    year = "2023",
}
```