Commit
•
625d5df
1
Parent(s):
5d2bf3f
Update README.md
Browse files
README.md
CHANGED
@@ -2,18 +2,18 @@
|
|
2 |
license: mit
|
3 |
language:
|
4 |
- en
|
5 |
-
tags:
|
6 |
-
- Mathematical Reasoning
|
7 |
datasets:
|
8 |
- akjindal53244/Arithmo-Data
|
|
|
|
|
9 |
---
|
10 |
|
11 |
-
**Arithmo2-Mistral-7B** model improves initially released Arithmo-Mistral-7B model on both GSM8K and MATH benchmarks. Specifically, there is **absolute** improvement of:
|
12 |
- +1.7% on GSM8K
|
13 |
- +3.0% on GSM8K PoT
|
14 |
- +1.9% on MATH
|
15 |
|
16 |
-
|
17 |
|
18 |
|
19 |
### Model Description
|
@@ -99,7 +99,7 @@ Plugging these values into the formula, we get:
|
|
99 |
The answer is: 55
|
100 |
```
|
101 |
|
102 |
-
Arithmo2-Mistral-7B is trained with
|
103 |
#### CoT Format (generate reasoning steps with answer):
|
104 |
```
|
105 |
Question: <question>
|
|
|
2 |
license: mit
|
3 |
language:
|
4 |
- en
|
|
|
|
|
5 |
datasets:
|
6 |
- akjindal53244/Arithmo-Data
|
7 |
+
tags:
|
8 |
+
- Mathematical Reasoning
|
9 |
---
|
10 |
|
11 |
+
**Arithmo2-Mistral-7B** model improves initially released [Arithmo-Mistral-7B](https://huggingface.co/akjindal53244/Arithmo-Mistral-7B) model on both GSM8K and MATH benchmarks. Specifically, there is **absolute** improvement of:
|
12 |
- +1.7% on GSM8K
|
13 |
- +3.0% on GSM8K PoT
|
14 |
- +1.9% on MATH
|
15 |
|
16 |
+
**This repo contains final merged model**. If you are interested in LoRA adapter, use [LoRA Adapter](https://huggingface.co/upaya07/Arithmo2-Mistral-7B-adapter) instead.
|
17 |
|
18 |
|
19 |
### Model Description
|
|
|
99 |
The answer is: 55
|
100 |
```
|
101 |
|
102 |
+
Arithmo2-Mistral-7B is trained with same format as [Arithmo-Mistral-7B](https://huggingface.co/akjindal53244/Arithmo-Mistral-7B):
|
103 |
#### CoT Format (generate reasoning steps with answer):
|
104 |
```
|
105 |
Question: <question>
|