Edit model card

Finetuning Overview:

Model Used: HuggingFaceH4/zephyr-7b-alpha
Dataset: meta-math/MetaMathQA

Dataset Insights:

The MetaMathQA dataset is a newly created dataset specifically designed for enhancing the mathematical reasoning capabilities of large language models (LLMs). It is built by bootstrapping mathematical questions and rewriting them from multiple perspectives, providing a comprehensive and challenging environment for LLMs to develop and refine their mathematical problem-solving skills.

Finetuning Details:

Using MonsterAPI's LLM finetuner, this finetuning:

  • Was conducted with efficiency and cost-effectiveness in mind.
  • Completed in a total duration of 10.9 hours for 0.5 epoch using an A6000 48GB GPU.
  • Costed $22.01 for the entire finetuning process.

Hyperparameters & Additional Details:

  • Epochs: 0.5
  • Total Finetuning Cost: $22.01
  • Model Path: HuggingFaceH4/zephyr-7b-alpha
  • Learning Rate: 0.0001
  • Data Split: 95% train 5% validation
  • Gradient Accumulation Steps: 4

Prompt Structure

Below is an instruction that describes a task. Write a response that appropriately completes the request.  


###Instruction:[query]


###Response:[response]

Training loss:

training loss


Benchmark Results:

GSM8K Accuracy

GSM8K is a dataset of 8.5K high quality linguistically diverse grade school math word problems, These problems take between 2 and 8 steps to solve, and solutions primarily involve performing a sequence of elementary calculations using basic arithmetic operations (+ − ×÷) to reach the final answer. A bright middle school student should be able to solve every problem. Its a industry wide used benchmark for testing an LLM for for multi-step mathematical reasoning.


license: apache-2.0

Downloads last month
5
Safetensors
Model size
7.24B params
Tensor type
BF16
·

Finetuned from

Dataset used to train monsterapi/zephyr-7b-alpha_metamathqa

Space using monsterapi/zephyr-7b-alpha_metamathqa 1