MM-Eureka: Exploring Visual Aha Moment with Rule-based Large-scale Reinforcement Learning
Abstract
We present MM-Eureka, a multimodal reasoning model that successfully extends large-scale <PRE_TAG>rule-based reinforcement learning (RL)</POST_TAG> to multimodal reasoning. While rule-based RL has shown remarkable success in improving LLMs' reasoning abilities in text domains, its application to multimodal settings has remained challenging. Our work reproduces key characteristics of text-based RL systems like DeepSeek-R1 in the multimodal space, including steady increases in accuracy reward and response length, and the emergence of reflection behaviors. We demonstrate that both instruction-tuned and pre-trained models can develop strong <PRE_TAG>multimodal reasoning capabilities</POST_TAG> through rule-based RL without supervised fine-tuning, showing superior data efficiency compared to alternative approaches. We open-source our complete pipeline to foster further research in this area. We release all our codes, models, data, etc. at https://github.com/ModalMinds/MM-EUREKA
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper