Datasets:

Modalities:
Tabular
Text
Formats:
json
Languages:
Chinese
ArXiv:
Libraries:
Datasets
pandas
License:
File size: 2,545 Bytes
bd996f7
 
6d35cae
 
 
 
 
 
bd996f7
6d35cae
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
eb98e61
ba61431
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
---
license: cc-by-4.0
language:
- zh
tags:
- mathematics
size_categories:
- 1K<n<10K
---
# CMATH
## Introduction
We present the Chinese Elementary School Math Word Problems (CMATH) dataset, comprising 1.7k elementary school-level math word problems with detailed annotations, source from actual Chinese workbooks and exams. This dataset aims to provide a benchmark tool for assessing the following question: to what grade level of elementary school math do the abilities of popular large language models (LLMs) correspond? We evaluate a variety of popular LLMs, including both commercial and open-source options, and discover that only GPT-4 achieves success (accuracy >= 60%) across all six elementary school grades, while other models falter at different grade levels.
Furthermore, we assess the robustness of LLMs by augmenting the original problems in the CMATH dataset with distracting information. Our findings reveal that GPT-4 is the sole model that maintains robustness, further distinguishing its performance from competing models. We anticipate that our CMATH dataset will expose limitations in LLMs' capabilities and promote their ongoing development and advancement.

## Datasets
### cmath_dev
Initial release of 600 examples from CMATH dataset, with 100 problems from each elementary school grade.  
We will release the remaining portion of the dataset by the end of the year.
#### Examples and Annotations
![Examples](assets/example1.png)

#### Evaluation Results
![Model Performance](assets/plot1.png)

### distractor
To assess the robustness of LLMs against "irrelevant" information, we manually created a small ``distractor dataset'' comprising 60 examples, 10 for each grade level. Each example consists of an original problem and five associated problems augmented with 1 ~ 5 piece(s) of irrelevant information which we refer to as distractor(s). 

#### Examples
![Examples](assets/example2.png)

#### Evaluation Results
![Model Performance](assets/plot2.png)

## Script
We provide a script `eval.py` that implements automated evaluation.

## License
MIT license

## Citation
```
@misc{wei2023cmath,
      title={CMATH: Can Your Language Model Pass Chinese Elementary School Math Test?}, 
      author={Tianwen Wei and Jian Luan and Wei Liu and Shuang Dong and Bin Wang},
      year={2023},
      eprint={2306.16636},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}
```


Visit our git [repository](https://github.com/XiaoMi/cmath) for more details.
You may also read our [paper](https://arxiv.org/abs/2306.16636).