Model Card for Hello-SimpleAI/chatgpt-qa-detector-roberta-chinese
This model is trained on question-answer
pairs of the filtered full-text from Hello-SimpleAI/HC3-Chinese.
More details refer to arxiv: 2301.07597 and Gtihub project Hello-SimpleAI/chatgpt-comparison-detection.
The base checkpoint is hfl/chinese-roberta-wwm-ext. We train it with all Hello-SimpleAI/HC3-Chinese data (without held-out) for 2 epochs.
( 2-epoch is consistent with the experiments in our paper.)
Citation
Checkout this papaer arxiv: 2301.07597
@article{guo-etal-2023-hc3,
title = "How Close is ChatGPT to Human Experts? Comparison Corpus, Evaluation, and Detection",
author = "Guo, Biyang and
Zhang, Xin and
Wang, Ziyuan and
Jiang, Minqi and
Nie, Jinran and
Ding, Yuxuan and
Yue, Jianwei and
Wu, Yupeng",
journal={arXiv preprint arxiv:2301.07597}
year = "2023",
}
- Downloads last month
- 26
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.