Commit
•
d809af3
1
Parent(s):
fc46879
Update README.md
Browse files
README.md
CHANGED
@@ -17,7 +17,7 @@ datasets:
|
|
17 |
|
18 |
## Model Details
|
19 |
|
20 |
-
**Arithmo2-7B** is improved version of [
|
21 |
|
22 |
### Model Description
|
23 |
|
@@ -30,7 +30,7 @@ datasets:
|
|
30 |
|
31 |
## Results
|
32 |
|
33 |
-
Arithmo2-7B is improved version of [
|
34 |
|
35 |
<table>
|
36 |
<thead>
|
@@ -102,7 +102,7 @@ Plugging these values into the formula, we get:
|
|
102 |
The answer is: 55
|
103 |
```
|
104 |
|
105 |
-
|
106 |
#### CoT Format (generate reasoning steps with answer):
|
107 |
```
|
108 |
Question: <question>
|
@@ -118,7 +118,7 @@ Answer:
|
|
118 |
```
|
119 |
It will perform best if queried in this way with your own script.
|
120 |
|
121 |
-
## Comparing
|
122 |
Results for all models except `Arithmo2-7B` are taken from [MetaMath](https://github.com/meta-math/MetaMath/blob/main/README.MD) repository.
|
123 |
|
124 |
| Model | GSM8k Pass@1 | MATH Pass@1 | Fine-tuning |
|
|
|
17 |
|
18 |
## Model Details
|
19 |
|
20 |
+
**Arithmo2-7B** is improved version of [Arithmo-Mistral-7B](https://huggingface.co/akjindal53244/Arithmo-Mistral-7B) model and is trained to reason and answer mathematical problems and is also capable of writing a Python program that upon execution prints answer to the question. We used [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1) as a base model and used **QLoRA to fine-tune it on a single GPU**.
|
21 |
|
22 |
### Model Description
|
23 |
|
|
|
30 |
|
31 |
## Results
|
32 |
|
33 |
+
Arithmo2-7B is improved version of [Arithmo-Mistral-7B](https://huggingface.co/akjindal53244/Arithmo-Mistral-7B) model and is competitive with full fine-tuned state-of-the-art 7B Mathematical Reasoning models. Refer to [Comparing Arithmo-Mistral-7B with other LLM models](https://github.com/akjindal53244/Arithmo-Mistral-7B/tree/master#comparing-arithmo-mistral-7b-with-other-llm-models) section for more details.
|
34 |
|
35 |
<table>
|
36 |
<thead>
|
|
|
102 |
The answer is: 55
|
103 |
```
|
104 |
|
105 |
+
Arithmo2-7B is trained with the following format:
|
106 |
#### CoT Format (generate reasoning steps with answer):
|
107 |
```
|
108 |
Question: <question>
|
|
|
118 |
```
|
119 |
It will perform best if queried in this way with your own script.
|
120 |
|
121 |
+
## Comparing Arithmo2-7B with other LLM models.
|
122 |
Results for all models except `Arithmo2-7B` are taken from [MetaMath](https://github.com/meta-math/MetaMath/blob/main/README.MD) repository.
|
123 |
|
124 |
| Model | GSM8k Pass@1 | MATH Pass@1 | Fine-tuning |
|