Text Generation
Transformers
Safetensors
English
mixtral
text-generation-inference
Inference Endpoints
File size: 3,467 Bytes
e120fcc
1ec90ea
 
e120fcc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a511458
e120fcc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
---
language:
- en
datasets:
- ToheartZhang/JiuZhang3.0-Corpus-PT-CoT
- ToheartZhang/JiuZhang3.0-Corpus-PT-Tool
- ToheartZhang/JiuZhang3.0-Corpus-SFT
---
<h1 align="center">
JiuZhang3.0: Efficiently Improving Mathematical
Reasoning by Training Small Data Synthesis Models
</h1>
<p align="center">
  <a href="https://arxiv.org/abs/2405.14365"><b>[Paper]</b></a><a href="https://github.com/RUCAIBox/JiuZhang3.0"><b>[GitHub]</b></a><a href="https://huggingface.co/collections/ToheartZhang/jiuzhang30-66508be8be5a61de47101655#/"><b>[Models]</b></a><a href="https://huggingface.co/collections/ToheartZhang/jiuzhang30-corpus-665092209525389ad7a2289a"><b>[Data]</b></a>
</p>

## Introduction
JiuZhang3.0 is a series of fine-tuned models for math reasoning continually pre-trained on corpus synthesized by our carefully trained small LLM.

## Experimental Results
For more evaluation results, please refer to the [Paper](https://arxiv.org/abs/2405.14365)

| Models                   | GSM8k | MATH | SVAMP | ASDiv | MAWPS | CARP | Avg.  |
|--------------------------|-------|------|-------|-------|-------|------|-------|
| GPT-4                    | 92.2  | 65.4 | 92.9  | 94.3  | 96.6  | 53.6 | 82.5  |
|**20B+ Models**||
| Llemma-34B               | 60.2  | 24.6 | 68.0  | 75.6  | 89.8  | 36.5 | 59.1  |
| Intern-Math-20B          | 64.9  | 27.4 | 74.9  | 79.6  | 94.4  | 42.3 | 63.9  |
| ChatGLM-Math-32B         | 82.6  | 40.6 | -     | -     | -     | -    | -     |
| MAmmoTH2-8x7B-Plus       | _86.4_| 47.0 | _90.0_| _92.2_| **97.0** | 45.8 | _76.4_ |
| [JiuZhang3.0-8x7B](https://huggingface.co/ToheartZhang/JiuZhang3.0-8x7B)   | **89.8** | **53.8** | **90.2** | **93.1** | _96.7_ | 52.3 | **79.3** |
|**7-8B Models**||
| Mistral-7B-MMIQC         | 75.0  | 34.2 | 73.5  | 82.1  | 90.1  | 36.5 | 65.2  |
| MetaMath-Mistral-7B      | 77.8  | 29.6 | 79.6  | 81.2  | 93.7  | 30.5 | 65.4  |
| Abel-7B-002              | 80.4  | 29.6 | 78.8  | 82.7  | 93.5  | 33.2 | 66.4  |
| WizardMath-7B-1.1        | 82.2  | 32.8 | 80.7  | 84.2  | 93.8  | 31.9 | 67.6  |
| Math-Shepherd-Mistral-7B | 84.3  | 34.4 | 82.9  | 82.8  | 92.5  | 32.9 | 68.3  |
| KPMath-DSMath-7B         | 83.9  | 48.8 | 81.5  | 88.9  | 94.8  | -    | -     |
| MAmmoTH2-7B-Plus         | 84.2  | 46.2 | _90.3_| 90.3  | _97.1_| 44.3 | 75.2  |
| MAmmoTH2-8B-Plus         | 84.4  | 41.2 | 89.9  | 89.9  | _97.1_| 44.8 | 74.6  |
| DeepSeekMath-7B-Instruct | 82.3  | 45.8 | 83.7  | 90.1  | 95.7  | 45.8 | 73.9  |
| DeepSeekMath-7B-RL       | 88.2  | 50.2 | 87.3  | 91.8  | 95.5  | **51.6** | 77.4  |
| [JiuZhang3.0-7B](https://huggingface.co/ToheartZhang/JiuZhang3.0-7B)     | **88.6** | **52.8** | **90.4** | **92.6** | **97.3** | _51.0_ | **78.8** |
| [JiuZhang3.0-8B](https://huggingface.co/ToheartZhang/JiuZhang3.0-8B)    | **88.6** | _51.0_ | 89.4  | **92.6** | _97.1_ | 50.9 | _78.3_ |

## Evaluation
### Natural Language Reasoning
```
## Question
{question}

## Solution
{solution}
```

### Tool Manipulation
```
## Question
{question}

## Code Solution
{solution}
```

## Citation
If you find this repository helpful, please consider citing our paper:

```
@article{zhou2024jiuzhang30,
      title={JiuZhang3.0: Efficiently Improving Mathematical Reasoning by Training Small Data Synthesis Models}, 
      author={Kun Zhou and Beichen Zhang and Jiapeng Wang and Zhipeng Chen and Wayne Xin Zhao and Jing Sha and Zhichao Sheng and Shijin Wang and Ji-Rong Wen},
      year={2024},
}
```