Model Card for Hello-SimpleAI/chatgpt-detector-roberta-chinese
This model is trained on the mix of full-text and splitted sentences of answer
s from Hello-SimpleAI/HC3-Chinese.
More details refer to arxiv: 2301.07597 and Gtihub project Hello-SimpleAI/chatgpt-comparison-detection.
The base checkpoint is hfl/chinese-roberta-wwm-ext. We train it with all Hello-SimpleAI/HC3-Chinese data (without held-out) for 2 epochs.
(2-epoch is consistent with the experiments in our paper.)
Citation
Checkout this papaer arxiv: 2301.07597
@article{guo-etal-2023-hc3,
title = "How Close is ChatGPT to Human Experts? Comparison Corpus, Evaluation, and Detection",
author = "Guo, Biyang and
Zhang, Xin and
Wang, Ziyuan and
Jiang, Minqi and
Nie, Jinran and
Ding, Yuxuan and
Yue, Jianwei and
Wu, Yupeng",
journal={arXiv preprint arxiv:2301.07597}
year = "2023",
}
- Downloads last month
- 393
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.