File size: 4,151 Bytes
5b9ab7f 8ffe199 2b62b13 8ffe199 2b62b13 5b9ab7f 72f91c6 4bf8555 72f91c6 c27cef8 72f91c6 4bf8555 72f91c6 4bf8555 72f91c6 4bf8555 342e482 c27cef8 385f319 c27cef8 385f319 4bf8555 385f319 c27cef8 4bf8555 342e482 4bf8555 72f91c6 4bf8555 72f91c6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 |
---
license: apache-2.0
configs:
- config_name: original-splits
data_files:
- split: train
path: original-splits/train-*
- split: validation
path: original-splits/validation-*
- split: test
path: original-splits/test-*
dataset_info:
- config_name: original-splits
features:
- name: id
dtype: string
- name: question
dtype: string
- name: chain
dtype: string
- name: result
dtype: string
- name: result_float
dtype: float64
- name: question_without_options
dtype: string
- name: options
struct:
- name: A
dtype: string
- name: B
dtype: string
- name: C
dtype: string
- name: D
dtype: string
- name: E
dtype: string
- name: annotated_formula
dtype: string
- name: linear_formula
dtype: string
- name: rationale
dtype: string
- name: category
dtype: string
splits:
- name: train
num_bytes: 25135321
num_examples: 20868
- name: validation
num_bytes: 3736735
num_examples: 3102
- name: test
num_bytes: 2431936
num_examples: 2029
download_size: 13918684
dataset_size: 31303992
---
# Dataset Card for "Calc-math_qa"
## Summary
This dataset is an instance of math_qa dataset, converted to a simple HTML-like language that can be easily parsed (e.g. by BeautifulSoup). The data contains 3 types of tags:
- gadget: A tag whose content is intended to be evaluated by calling an external tool (sympy-based calculator in this case)
- output: An output of the external tool
- result: The final answer of the mathematical problem (correct option)
## Supported Tasks
The dataset is intended for training Chain-of-Thought reasoning **models able to use external tools** to enhance the factuality of their responses.
This dataset presents in-context scenarios where models can outsource the computations in the reasoning chain to a calculator.
## Construction Process
We took the original math_qa dataset, parsed the nested formulas, linearized them into a sequence (chain) of operations, and replace all advanced
function calls (such as `circle_area`) with explicit elementary operations. We evaluate all the steps in each example and filter out examples if their
evaluation does not match the answer selected as correct in the data with a 5% tolerance. The sequence of steps is then saved in HTML-like language
in the `chain` column. We keep the original columns in the dataset for convenience.
You can read more information about this process in our [technical report](https://arxiv.org/abs/2305.15017).
## Content and Data splits
Content and splits correspond to the original math_qa dataset.
See [mathqa HF dataset](https://huggingface.co/datasets/math_qa) and [official website](https://math-qa.github.io/) for more info.
Columns:
- `question` - th description of a mathematical problem in natural language
- `options` - dictionary with choices 'A' to 'E' as possible solutions
- `chain` - solution in the form of step-by-step calculations encoded in simple html-like language. computed from `annotated_formula` column
- `result` - the correct option
- `result_float` - the result converted to a float
- `annotated_formula` - human-annotated nested expression that (approximately) evaluates to the selected correct answer
- `linear_formula` - same as `annotated_formula`, but linearized by original math_qa authors
- `rationale` - human-annotated free-text reasoning that leads to the correct answer
- `index` - index of the example in the original math_qa dataset
## Licence
Apache 2.0, consistently with the original dataset.
## Cite
If you use this version of dataset in research, please cite the [original MathQA paper](https://arxiv.org/abs/1905.13319), and also [our technical report](https://arxiv.org/abs/2305.15017) as follows:
```bibtex
@article{kadlcik2023calcx,
title={Calc-X: Enriching Arithmetical Chain-of-Thoughts Datasets by Interaction with Symbolic Systems},
author={Marek Kadlčík and Michal Štefánik},
year={2023},
eprint={2305.15017},
archivePrefix={arXiv},
primaryClass={cs.LG}
}
``` |