Update README.md (#11)
Browse files- Update README.md (e5103930e082d6cb05c400eeaff765a84e2eddd8)
Co-authored-by: haipengluo <haipeng1@users.noreply.huggingface.co>
README.md
CHANGED
@@ -6,9 +6,8 @@ license: llama2
|
|
6 |
## WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct (RLEIF)
|
7 |
|
8 |
|
9 |
-
|
10 |
<p align="center">
|
11 |
-
π€ <a href="https://huggingface.co/WizardLM" target="_blank">HF Repo</a>
|
12 |
</p>
|
13 |
<p align="center">
|
14 |
π Join our <a href="https://discord.gg/VZjjHtWrKs" target="_blank">Discord</a>
|
@@ -24,13 +23,13 @@ license: llama2
|
|
24 |
| WizardCoder-1B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardCoder-1B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2306.08568" target="_blank">[WizardCoder]</a> | 23.8 |28.6 | -- | <a href="https://huggingface.co/spaces/bigcode/bigcode-model-license-agreement" target="_blank">OpenRAIL-M</a> |
|
25 |
|
26 |
|
27 |
-
|
28 |
| Model | Checkpoint | Paper | GSM8k | MATH |Online Demo| License|
|
29 |
| ----- |------| ---- |------|-------| ----- | ----- |
|
30 |
| WizardMath-70B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-70B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **81.6** | **22.7** |[Demo](http://47.103.63.15:50083/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a> |
|
31 |
| WizardMath-13B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-13B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **63.9** | **14.0** |[Demo](http://47.103.63.15:50082/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a> |
|
32 |
| WizardMath-7B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-7B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **54.9** | **10.7** | [Demo](http://47.103.63.15:50080/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a>|
|
33 |
|
|
|
34 |
|
35 |
<font size=4>
|
36 |
|
@@ -50,7 +49,13 @@ license: llama2
|
|
50 |
|
51 |
**Discord**: https://discord.gg/VZjjHtWrKs
|
52 |
|
|
|
|
|
|
|
53 |
|
|
|
|
|
|
|
54 |
|
55 |
β<b>Note for model system prompts usage:</b>
|
56 |
|
@@ -77,3 +82,21 @@ Recently, there have been clear changes in the open-source policy and regulation
|
|
77 |
Despite this, we have still worked hard to obtain opening the weights of the model first, but the data involves stricter auditing and is in review with our legal team .
|
78 |
Our researchers have no authority to publicly release them without authorization.
|
79 |
Thank you for your understanding.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
6 |
## WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct (RLEIF)
|
7 |
|
8 |
|
|
|
9 |
<p align="center">
|
10 |
+
π€ <a href="https://huggingface.co/WizardLM" target="_blank">HF Repo</a> β’π± <a href="https://github.com/nlpxucan/WizardLM" target="_blank">Github Repo</a> β’ π¦ <a href="https://twitter.com/WizardLM_AI" target="_blank">Twitter</a> β’ π <a href="https://arxiv.org/abs/2304.12244" target="_blank">[WizardLM]</a> β’ π <a href="https://arxiv.org/abs/2306.08568" target="_blank">[WizardCoder]</a> β’ π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a> <br>
|
11 |
</p>
|
12 |
<p align="center">
|
13 |
π Join our <a href="https://discord.gg/VZjjHtWrKs" target="_blank">Discord</a>
|
|
|
23 |
| WizardCoder-1B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardCoder-1B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2306.08568" target="_blank">[WizardCoder]</a> | 23.8 |28.6 | -- | <a href="https://huggingface.co/spaces/bigcode/bigcode-model-license-agreement" target="_blank">OpenRAIL-M</a> |
|
24 |
|
25 |
|
|
|
26 |
| Model | Checkpoint | Paper | GSM8k | MATH |Online Demo| License|
|
27 |
| ----- |------| ---- |------|-------| ----- | ----- |
|
28 |
| WizardMath-70B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-70B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **81.6** | **22.7** |[Demo](http://47.103.63.15:50083/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a> |
|
29 |
| WizardMath-13B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-13B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **63.9** | **14.0** |[Demo](http://47.103.63.15:50082/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a> |
|
30 |
| WizardMath-7B-V1.0 | π€ <a href="https://huggingface.co/WizardLM/WizardMath-7B-V1.0" target="_blank">HF Link</a> | π <a href="https://arxiv.org/abs/2308.09583" target="_blank">[WizardMath]</a>| **54.9** | **10.7** | [Demo](http://47.103.63.15:50080/)| <a href="https://ai.meta.com/resources/models-and-libraries/llama-downloads/" target="_blank">Llama 2 </a>|
|
31 |
|
32 |
+
|
33 |
|
34 |
<font size=4>
|
35 |
|
|
|
49 |
|
50 |
**Discord**: https://discord.gg/VZjjHtWrKs
|
51 |
|
52 |
+
## Comparing WizardMath-V1.0 with Other LLMs.
|
53 |
+
|
54 |
+
π₯ The following figure shows that our **WizardMath-70B-V1.0 attains the fifth position in this benchmark**, surpassing ChatGPT (81.6 vs. 80.8) , Claude Instant (81.6 vs. 80.9), PaLM 2 540B (81.6 vs. 80.7).
|
55 |
|
56 |
+
<p align="center" width="100%">
|
57 |
+
<a ><img src="https://raw.githubusercontent.com/nlpxucan/WizardLM/main/WizardMath/images/wizardmath_gsm8k.png" alt="WizardMath" style="width: 96%; min-width: 300px; display: block; margin: auto;"></a>
|
58 |
+
</p>
|
59 |
|
60 |
β<b>Note for model system prompts usage:</b>
|
61 |
|
|
|
82 |
Despite this, we have still worked hard to obtain opening the weights of the model first, but the data involves stricter auditing and is in review with our legal team .
|
83 |
Our researchers have no authority to publicly release them without authorization.
|
84 |
Thank you for your understanding.
|
85 |
+
|
86 |
+
|
87 |
+
## Inference Demo Script
|
88 |
+
|
89 |
+
We provide the inference demo code [here](https://github.com/nlpxucan/WizardLM/tree/main/demo).
|
90 |
+
|
91 |
+
## Citation
|
92 |
+
|
93 |
+
Please cite the repo if you use the data, method or code in this repo.
|
94 |
+
|
95 |
+
```
|
96 |
+
@article{luo2023wizardmath,
|
97 |
+
title={WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct},
|
98 |
+
author={Luo, Haipeng and Sun, Qingfeng and Xu, Can and Zhao, Pu and Lou, Jianguang and Tao, Chongyang and Geng, Xiubo and Lin, Qingwei and Chen, Shifeng and Zhang, Dongmei},
|
99 |
+
journal={arXiv preprint arXiv:2308.09583},
|
100 |
+
year={2023}
|
101 |
+
}
|
102 |
+
```
|