Zangs3011 commited on
Commit
43e56b1
1 Parent(s): 3f32599

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +69 -0
README.md ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - meta-math
5
+ - code
6
+ - instruct
7
+ - Zephyr-7B-Alpha
8
+ datasets:
9
+ - meta-math/MetaMathQA
10
+ base_model: HuggingFaceH4/zephyr-7b-alpha
11
+ license: apache-2.0
12
+ ---
13
+
14
+ ### Finetuning Overview:
15
+
16
+ **Model Used:** HuggingFaceH4/zephyr-7b-alpha
17
+ **Dataset:** meta-math/MetaMathQA
18
+
19
+ #### Dataset Insights:
20
+
21
+ The MetaMathQA dataset is a newly created dataset specifically designed for enhancing the mathematical reasoning capabilities of large language models (LLMs). It is built by bootstrapping mathematical questions and rewriting them from multiple perspectives, providing a comprehensive and challenging environment for LLMs to develop and refine their mathematical problem-solving skills.
22
+
23
+ #### Finetuning Details:
24
+
25
+ Using [MonsterAPI](https://monsterapi.ai)'s [LLM finetuner](https://docs.monsterapi.ai/fine-tune-a-large-language-model-llm), this finetuning:
26
+
27
+ - Was conducted with efficiency and cost-effectiveness in mind.
28
+ - Completed in a total duration of 10.9 hours for 0.5 epoch using an A6000 48GB GPU.
29
+ - Costed `$22.01` for the entire finetuning process.
30
+
31
+ #### Hyperparameters & Additional Details:
32
+
33
+ - **Epochs:** 0.5
34
+ - **Total Finetuning Cost:** $22.01
35
+ - **Model Path:** HuggingFaceH4/zephyr-7b-alpha
36
+ - **Learning Rate:** 0.0002
37
+ - **Data Split:** 90% train 10% validation
38
+ - **Gradient Accumulation Steps:** 4
39
+
40
+ ---
41
+ Prompt Structure
42
+
43
+ ```
44
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
45
+
46
+
47
+ ###Instruction:[query]
48
+
49
+
50
+ ###Response:[response]
51
+ ```
52
+
53
+ ---
54
+ ### Training loss:
55
+ ![training loss](zephyr-mmqa-1.png "Training loss")
56
+
57
+
58
+
59
+ ---
60
+ ### Benchmark Results:
61
+
62
+ ![GSM8K Accuracy ](benchmark.png "GSM8K Accuracy")
63
+
64
+ GSM8K is a dataset of 8.5K high quality linguistically diverse grade school math word problems, These problems take between 2 and 8 steps to solve, and solutions primarily involve performing a sequence of elementary calculations using basic arithmetic operations (+ − ×÷) to reach the final answer. A bright middle school student should be able to solve every problem. Its a industry wide used benchmark for testing an LLM for for multi-step mathematical reasoning.
65
+
66
+
67
+
68
+ ---
69
+ license: apache-2.0