Update README.md
Browse files
README.md
CHANGED
@@ -8,6 +8,13 @@ see our paper in https://arxiv.org/abs/2309.12284
|
|
8 |
View the project page:
|
9 |
https://meta-math.github.io/
|
10 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
## Model Details
|
12 |
|
13 |
MetaMath-Mistral-7B is fully fine-tuned on the MetaMathQA datasets and based on the powerful Mistral-7B model. It is glad to see using MetaMathQA datasets and change the base model from llama-2-7B to Mistral-7b can boost the GSM8K performance from 66.5 to **77.7**.
|
|
|
8 |
View the project page:
|
9 |
https://meta-math.github.io/
|
10 |
|
11 |
+
## Note
|
12 |
+
|
13 |
+
All MetaMathQA data are augmented from the training sets of GSM8K and MATH.
|
14 |
+
<span style="color:red"><b>None of the augmented data is from the testing set.</b></span>
|
15 |
+
|
16 |
+
You can check the `original_question` in `meta-math/MetaMathQA`, each item is from the GSM8K or MATH train set.
|
17 |
+
|
18 |
## Model Details
|
19 |
|
20 |
MetaMath-Mistral-7B is fully fine-tuned on the MetaMathQA datasets and based on the powerful Mistral-7B model. It is glad to see using MetaMathQA datasets and change the base model from llama-2-7B to Mistral-7b can boost the GSM8K performance from 66.5 to **77.7**.
|