llemma_7b_muinstruct_camelmath is an instruction-following finetune of Llemma 7B, trained on the μInstruct and camel-ai/math datasets.

Input Formatting

Format input queries as follows:

input_text = f"Input:{input}\n\nResponse:"

Note that due to an error during training, this model's end-of-sequence token ID is 0 instead of the 2 which is standard for Llama-2 based models. Inference APIs should handle this automatically by reading this repo's config.json, but be aware of this difference if you are doing token surgery.

Evals

llemma_7b_muinstruct_camelmath compares favorably to other 7B parameter models on the Hungarian Math Exam. It surpasses the few-shot performance of Llemma 7B whilst being the strongest Llama-2 7B based model.

Model Exam Score
Code Llama 7B (few-shot) 8%
MetaMath 7B 20%
MAmmoTH 7B 17%
MAmmoTH Coder 7B 11%
Llemma 7B (few-shot) 23%
Llemma_7B_muinstruct_camelmath 25%
- -
Mistral 7B (few-shot) 22%
MetaMath Mistral 7B 29%
OpenChat 3.5 37%
Downloads last month
337
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Datasets used to train EleutherAI/llemma_7b_muinstruct_camelmath