General Preference Modeling with Preference Representations for Aligning Language Models (https://arxiv.org/abs/2410.02197)
SPPO-Llama-3-8B-Instruct-GPM-2B
This model was developed using SPPO at iteration 3 and the General Preference representation Model (GPM) (specifically, using GPM-Gemma-2B), based on the meta-llama/Meta-Llama-3-8B-Instruct architecture as starting point. We utilized the prompt sets from the openbmb/UltraFeedback dataset, splited to 3 parts for 3 iterations by snorkelai/Snorkel-Mistral-PairRM-DPO-Dataset. All responses used are synthetic.
Links to Other Models
Model Description
- Model type: A 8B parameter GPT-like model fine-tuned on synthetic datasets.
- Language(s) (NLP): Primarily English
- License: Apache-2.0
- Finetuned from model: meta-llama/Meta-Llama-3-8B-Instruct
AlpacaEval Leaderboard Evaluation Results
Model | LC. Win Rate | Win Rate | Avg. Length |
---|---|---|---|
SPPO-Llama-3-8B-Instruct-GPM-2B | 35.30 | 45.44 | 2490 |
Open LLM Leaderboard Evaluation Results
Results are reported by using lm-evaluation-harness v0.4.1
arc_challenge | truthfulqa_mc2 | winogrande | gsm8k | hellaswag | mmlu | average | |
---|---|---|---|---|---|---|---|
SPPO-Llama-3-8B-Instruct-GPM-2B | 62.03 | 52.95 | 76.56 | 75.36 | 78.57 | 65.66 | 68.52 |
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-07
- eta: 1000
- per_device_train_batch_size: 8
- gradient_accumulation_steps: 1
- seed: 42
- distributed_type: deepspeed_zero3
- num_devices: 8
- optimizer: RMSProp
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- num_train_epochs: 6.0 (stop at epoch=1.0)
Citation
@article{zhang2024general,
title={General Preference Modeling with Preference Representations for Aligning Language Models},
author={Zhang, Yifan and Zhang, Ge and Wu, Yue and Xu, Kangping and Gu, Quanquan},
journal={arXiv preprint arXiv:2410.02197},
year={2024}
}
- Downloads last month
- 16
Dataset used to train general-preference/SPPO-Llama-3-8B-Instruct-GPM-2B
Evaluation results
- strict accuracy on IFEval (0-Shot)Open LLM Leaderboard60.240
- normalized accuracy on BBH (3-Shot)Open LLM Leaderboard27.890
- exact match on MATH Lvl 5 (4-Shot)Open LLM Leaderboard8.010
- acc_norm on GPQA (0-shot)Open LLM Leaderboard1.230
- acc_norm on MuSR (0-shot)Open LLM Leaderboard3.190
- accuracy on MMLU-PRO (5-shot)test set Open LLM Leaderboard29.530