Edit model card

Model Card for RLHF-V

Project Page | GitHub | Demo | Paper

RLHF-V is an open-source multimodal large language model with the lowest hallucination rate on both long-form instructions and short-form questions.

RLHF-V is trained on RLHF-V-Dataset, which contains fine-grained segment-level human corrections on diverse instructions. The base model is trained on UniMM-Chat, which is a high-quality knowledge-intensive SFT dataset. We introduce a new method Dense Direct Preference Optimization (DDPO) that can make better use of the fine-grained annotations.

For more details, please refer to our paper.

Illustration of the RLHF-V frmework

Model Details

Model Description

Model Sources

Performance

Low hallucination rate while being informative:

fig2

More resistant to over-generalization, even compared to GPT-4V:

img

Citation

If you find RLHF-V is useful in your work, please cite it with:

@article{2023rlhf-v,
  author      = {Tianyu Yu and Yuan Yao and Haoye Zhang and Taiwen He and Yifeng Han and Ganqu Cui and Jinyi Hu and Zhiyuan Liu and Hai-Tao Zheng and Maosong Sun and Tat-Seng Chua},
  title       = {RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback},
  journal      = {arxiv},
  year         = {2023},
}
Downloads last month
6

Datasets used to train openbmb/RLHF-V