Update README.md
Browse files
README.md
CHANGED
|
@@ -185,7 +185,16 @@ print("RM Scores:", scores)
|
|
| 185 |
Please refer to [https://github.com/LARK-AI-Lab/CodeScaler](https://github.com/LARK-AI-Lab/CodeScaler) for rl training details.
|
| 186 |
|
| 187 |
## Citation
|
|
|
|
| 188 |
```
|
| 189 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 190 |
```
|
| 191 |
|
|
|
|
| 185 |
Please refer to [https://github.com/LARK-AI-Lab/CodeScaler](https://github.com/LARK-AI-Lab/CodeScaler) for rl training details.
|
| 186 |
|
| 187 |
## Citation
|
| 188 |
+
If you find our work helpful, please consider citing:
|
| 189 |
```
|
| 190 |
+
@misc{zhu2026codescalerscalingcodellm,
|
| 191 |
+
title={CodeScaler: Scaling Code LLM Training and Test-Time Inference via Execution-Free Reward Models},
|
| 192 |
+
author={Xiao Zhu and Xinyu Zhou and Boyu Zhu and Hanxu Hu and Mingzhe Du and Haotian Zhang and Huiming Wang and Zhijiang Guo},
|
| 193 |
+
year={2026},
|
| 194 |
+
eprint={2602.17684},
|
| 195 |
+
archivePrefix={arXiv},
|
| 196 |
+
primaryClass={cs.LG},
|
| 197 |
+
url={https://arxiv.org/abs/2602.17684},
|
| 198 |
+
}
|
| 199 |
```
|
| 200 |
|