akjindal53244 commited on
Commit
d809af3
1 Parent(s): fc46879

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -17,7 +17,7 @@ datasets:
17
 
18
  ## Model Details
19
 
20
- **Arithmo2-7B** is improved version of [Arithmi-Mistral-7B](https://huggingface.co/akjindal53244/Arithmo-Mistral-7B) model and is trained to reason and answer mathematical problems and is also capable of writing a Python program that upon execution prints answer to the question. We used [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1) as a base model and used **QLoRA to fine-tune it on a single GPU**.
21
 
22
  ### Model Description
23
 
@@ -30,7 +30,7 @@ datasets:
30
 
31
  ## Results
32
 
33
- Arithmo2-7B is improved version of [Arithmi-Mistral-7B](https://huggingface.co/akjindal53244/Arithmo-Mistral-7B) model and is competitive with full fine-tuned state-of-the-art 7B Mathematical Reasoning models. Refer to [Comparing Arithmo-Mistral-7B with other LLM models](https://github.com/akjindal53244/Arithmo-Mistral-7B/tree/master#comparing-arithmo-mistral-7b-with-other-llm-models) section for more details.
34
 
35
  <table>
36
  <thead>
@@ -102,7 +102,7 @@ Plugging these values into the formula, we get:
102
  The answer is: 55
103
  ```
104
 
105
- Arithmo-Mistral-7B is trained with the following format:
106
  #### CoT Format (generate reasoning steps with answer):
107
  ```
108
  Question: <question>
@@ -118,7 +118,7 @@ Answer:
118
  ```
119
  It will perform best if queried in this way with your own script.
120
 
121
- ## Comparing Arithmo-Mistral-7B with other LLM models.
122
  Results for all models except `Arithmo2-7B` are taken from [MetaMath](https://github.com/meta-math/MetaMath/blob/main/README.MD) repository.
123
 
124
  | Model | GSM8k Pass@1 | MATH Pass@1 | Fine-tuning |
 
17
 
18
  ## Model Details
19
 
20
+ **Arithmo2-7B** is improved version of [Arithmo-Mistral-7B](https://huggingface.co/akjindal53244/Arithmo-Mistral-7B) model and is trained to reason and answer mathematical problems and is also capable of writing a Python program that upon execution prints answer to the question. We used [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1) as a base model and used **QLoRA to fine-tune it on a single GPU**.
21
 
22
  ### Model Description
23
 
 
30
 
31
  ## Results
32
 
33
+ Arithmo2-7B is improved version of [Arithmo-Mistral-7B](https://huggingface.co/akjindal53244/Arithmo-Mistral-7B) model and is competitive with full fine-tuned state-of-the-art 7B Mathematical Reasoning models. Refer to [Comparing Arithmo-Mistral-7B with other LLM models](https://github.com/akjindal53244/Arithmo-Mistral-7B/tree/master#comparing-arithmo-mistral-7b-with-other-llm-models) section for more details.
34
 
35
  <table>
36
  <thead>
 
102
  The answer is: 55
103
  ```
104
 
105
+ Arithmo2-7B is trained with the following format:
106
  #### CoT Format (generate reasoning steps with answer):
107
  ```
108
  Question: <question>
 
118
  ```
119
  It will perform best if queried in this way with your own script.
120
 
121
+ ## Comparing Arithmo2-7B with other LLM models.
122
  Results for all models except `Arithmo2-7B` are taken from [MetaMath](https://github.com/meta-math/MetaMath/blob/main/README.MD) repository.
123
 
124
  | Model | GSM8k Pass@1 | MATH Pass@1 | Fine-tuning |