Update README.md
Browse files
README.md
CHANGED
|
@@ -9,60 +9,22 @@ tags:
|
|
| 9 |
licence: license
|
| 10 |
---
|
| 11 |
|
| 12 |
-
|
| 13 |
|
| 14 |
-
|
| 15 |
-
It has been trained using [TRL](https://github.com/huggingface/trl).
|
| 16 |
|
| 17 |
-
|
| 18 |
|
| 19 |
-
|
| 20 |
-
from transformers import pipeline
|
| 21 |
|
| 22 |
-
|
| 23 |
-
generator = pipeline("text-generation", model="None", device="cuda")
|
| 24 |
-
output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
|
| 25 |
-
print(output["generated_text"])
|
| 26 |
-
```
|
| 27 |
|
| 28 |
-
|
| 29 |
|
| 30 |
-
|
| 31 |
|
| 32 |
-
|
| 33 |
-
This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300).
|
| 34 |
|
| 35 |
### Framework versions
|
| 36 |
|
| 37 |
-
-
|
| 38 |
-
- Transformers: 4.53.0
|
| 39 |
-
- Pytorch: 2.8.0.dev20250605+cu128
|
| 40 |
-
- Datasets: 3.6.0
|
| 41 |
-
- Tokenizers: 0.21.2
|
| 42 |
-
|
| 43 |
-
## Citations
|
| 44 |
-
|
| 45 |
-
Cite GRPO as:
|
| 46 |
-
|
| 47 |
-
```bibtex
|
| 48 |
-
@article{zhihong2024deepseekmath,
|
| 49 |
-
title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}},
|
| 50 |
-
author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo},
|
| 51 |
-
year = 2024,
|
| 52 |
-
eprint = {arXiv:2402.03300},
|
| 53 |
-
}
|
| 54 |
-
|
| 55 |
-
```
|
| 56 |
-
|
| 57 |
-
Cite TRL as:
|
| 58 |
-
|
| 59 |
-
```bibtex
|
| 60 |
-
@misc{vonwerra2022trl,
|
| 61 |
-
title = {{TRL: Transformer Reinforcement Learning}},
|
| 62 |
-
author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
|
| 63 |
-
year = 2020,
|
| 64 |
-
journal = {GitHub repository},
|
| 65 |
-
publisher = {GitHub},
|
| 66 |
-
howpublished = {\url{https://github.com/huggingface/trl}}
|
| 67 |
-
}
|
| 68 |
-
```
|
|
|
|
| 9 |
licence: license
|
| 10 |
---
|
| 11 |
|
| 12 |
+
## Model Card for Qwen3-14B-ZeroGPT-beta-step-150
|
| 13 |
|
| 14 |
+
### Model Details
|
|
|
|
| 15 |
|
| 16 |
+
This model was fine tuned with GRPO using an inverted score from [trentmkelly/zerogpt_distil](https://huggingface.co/trentmkelly/zerogpt_distil) as the reward function.
|
| 17 |
|
| 18 |
+
This model is extremely effective at evading the ZeroGPT classifier specifically, scoring an average of 6.8% AI across 8,080 tested essays. Out of these, 8,017 essays were successful evasions and 63 were failed evasions.
|
|
|
|
| 19 |
|
| 20 |
+
This model performs unpredictably on other AI detection tools.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 21 |
|
| 22 |
+
### System Prompt
|
| 23 |
|
| 24 |
+
The system prompt used during training was `/no_think\nYou are an essay writer. Write like a human. You will be graded on how human you sound, so try to avoid sounding like AI. Your essay should be 5 paragraphs long.`
|
| 25 |
|
| 26 |
+
Thinking mode hasn't been tested nor have other variations from this prompt. Variations will probably affect how the model performs versus the real classifier.
|
|
|
|
| 27 |
|
| 28 |
### Framework versions
|
| 29 |
|
| 30 |
+
- PEFT 0.15.2
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|