--- pipeline_tag: text-generation --- ## OmniLMM 12B **OmniLMM-12B** is the most capable version. The model is built based on [EVA02-5B](https://github.com/baaivision/EVA/tree/master/EVA-CLIP) and [Zephyr-7B-β](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta), connected with a perceiver resampler layer, and trained on multimodal data in a curriculum fashion. The model has three notable features: - 🔥 **Strong Performance.** OmniLMM-12B achieves **leading performance** among models with comparable sizes, surpassing established LMMs on multiple benchmarks (including MME, MMBench, SEED-Bench, etc). The model also **supports OCR capability** and endows **rich multimodal world knowledge**. - 🏆 **Trustworthy Behavior.** LMMs are known for suffering from hallucination, often generating text that is not factually grounded in images (e.g., faithfully describing non-existing objects in images). OmniLMM-12B is **the first state-of-the-art open-source LMM aligned via multimodal RLHF for trustworthy behavior** (using our recent [RLHF-V](https://rlhf-v.github.io/) technique) and **ranked #1** among open-source models on [MMHal-Bench](https://huggingface.co/datasets/Shengcao1006/MMHal-Bench). - 🕹 **Real-time Multimodal Interaction.** We combine the OmniLMM-12B and GPT-3.5 into a **real-time multimodal interactive assistant**. The assistant accepts video streams from the camera and speech streams from the microphone and emits speech output. While still primary, we find the model can **replicate some of the fun cases shown in the Gemini Demo video, without any video edition**.
Model | Size | MME | MMMU val | MMHal-Bench | SeedBench-I | LLaVA Bench W | MathVista | MMB dev (en) |
---|---|---|---|---|---|---|---|---|
GPT-4V † | - | 1409 | 56.8 | 3.53 / 70.8 | 71.6 | 93.1 | 47.8 | 75.1 |
Qwen-VL-Plus † | - | 1681 | 45.2 | - | 65.7 | 73.7 | 36.0 | 66.2 |
Yi-VL 6B | 6.7B | - | 39.1 | - | 66.1 | 39.9 | 28.0 | 68.2 |
CogVLM | 17.4B | 1438 | 32.1 | 2.68 / 52.1 | 68.8 | 73.9 | 34.7 | 63.7 |
Qwen-VL-Chat | 9.6B | 1488 | 35.9 | 2.93 / 59.4 | 64.8 | 67.7 | 33.8 | 60.6 |
LLaVA 1.5 | 13.6B | 1531 | 36.4 | 2.71 / 51.0 | 68.1 | 64.6 | 26.4 | 68.2 |
OmniLMM-12B | 11.6B | 1637 | 40.7 | 3.45 / 68.8 | 71.1 | 72.0 | 34.9 | 71.6 |