AI & ML interests

None defined yet.

Recent Activity

Data and models accompanying the paper When To Solve, When To Verify: Compute-Optimal Problem Solving and Generative Verification for LLM Reasoning, containing:

  • Finetuned generative verifiers (i.e., GenRM-FT) for math reasoning.
  • Synthetic verification data generated by GPT-4o for math reasoning to train your own generative verifiers.
  • Solutions and verifications generated by various models for math and science reasoning.

MATH Dataset

We use Llama-3.1-8B-Instruct and Qwen-2.5-7B-Instruct to generate solutions for problems in the training split of the MATH dataset. Then, we use GPT-4o to verify these solutions. We filter out the verifications whose verdict doesn't match the ground-truth correctness of the solution, and balance the dataset to have equal 'yes' and 'no' verifications in the dataset. This results in these datasets:

Training data for GenRM-FT

We fine-tune the two models on their respective datasets using LoRA, resulting in these fine-tuned GenRMs:

Finetuned Verifiers:

You can follow this example of how to do inference with these models.

We use these generative verifiers (without fine-tuning in the case of Llama-3.3-70B-Instruct) on solutions from the MATH test set to obtain this data, which we analyse in the paper:

Solutions and Verifications for Test-set

AIME25

Solutions and Verifications

GPQA

Solutions and Verifications