adalaw's picture
Update README.md
1754e42 verified
|
raw
history blame
1.13 kB
metadata
datasets:
  - TIGER-Lab/MathInstruct

Introduction

The model is trained with Masked Thought Fine-Tuning (MFT), a simple variant of standard Supervised Fine-Tuning (SFT). You can refer to our code and paper below.

Links

Results

We test it with the Hybrid decoding scripts provided in MAmmoTH.

Model GSM8K MATH
adalaw/MAmmoTH-7B-Mistral-MFT 77.10 41.2
TIGER-Lab/MAmmoTH-7B-Mistral-SFT 75.00 40.0