akjindal53244 commited on
Commit
625d5df
1 Parent(s): 5d2bf3f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -2,18 +2,18 @@
2
  license: mit
3
  language:
4
  - en
5
- tags:
6
- - Mathematical Reasoning
7
  datasets:
8
  - akjindal53244/Arithmo-Data
 
 
9
  ---
10
 
11
- **Arithmo2-Mistral-7B** model improves initially released Arithmo-Mistral-7B model on both GSM8K and MATH benchmarks. Specifically, there is **absolute** improvement of:
12
  - +1.7% on GSM8K
13
  - +3.0% on GSM8K PoT
14
  - +1.9% on MATH
15
 
16
- We release both [merged model](https://huggingface.co/upaya07/Arithmo2-Mistral-7B) and [LoRA Adapter](https://huggingface.co/upaya07/Arithmo2-Mistral-7B-adapter).
17
 
18
 
19
  ### Model Description
@@ -99,7 +99,7 @@ Plugging these values into the formula, we get:
99
  The answer is: 55
100
  ```
101
 
102
- Arithmo2-Mistral-7B is trained with the following format:
103
  #### CoT Format (generate reasoning steps with answer):
104
  ```
105
  Question: <question>
 
2
  license: mit
3
  language:
4
  - en
 
 
5
  datasets:
6
  - akjindal53244/Arithmo-Data
7
+ tags:
8
+ - Mathematical Reasoning
9
  ---
10
 
11
+ **Arithmo2-Mistral-7B** model improves initially released [Arithmo-Mistral-7B](https://huggingface.co/akjindal53244/Arithmo-Mistral-7B) model on both GSM8K and MATH benchmarks. Specifically, there is **absolute** improvement of:
12
  - +1.7% on GSM8K
13
  - +3.0% on GSM8K PoT
14
  - +1.9% on MATH
15
 
16
+ **This repo contains final merged model**. If you are interested in LoRA adapter, use [LoRA Adapter](https://huggingface.co/upaya07/Arithmo2-Mistral-7B-adapter) instead.
17
 
18
 
19
  ### Model Description
 
99
  The answer is: 55
100
  ```
101
 
102
+ Arithmo2-Mistral-7B is trained with same format as [Arithmo-Mistral-7B](https://huggingface.co/akjindal53244/Arithmo-Mistral-7B):
103
  #### CoT Format (generate reasoning steps with answer):
104
  ```
105
  Question: <question>