Papers
arxiv:2402.12226

AnyGPT: Unified Multimodal LLM with Discrete Sequence Modeling

Published on Feb 19
· Featured in Daily Papers on Feb 20
Authors:
,
,
,
,
,
,
Jie Fu ,
,
,

Abstract

We introduce AnyGPT, an any-to-any multimodal language model that utilizes discrete representations for the unified processing of various modalities, including speech, text, images, and music. AnyGPT can be trained stably without any alterations to the current large language model (LLM) architecture or training paradigms. Instead, it relies exclusively on data-level preprocessing, facilitating the seamless integration of new modalities into LLMs, akin to the incorporation of new languages. We build a multimodal text-centric dataset for multimodal alignment pre-training. Utilizing generative models, we synthesize the first large-scale any-to-any multimodal instruction dataset. It consists of 108k samples of multi-turn conversations that intricately interweave various modalities, thus equipping the model to handle arbitrary combinations of multimodal inputs and outputs. Experimental results demonstrate that AnyGPT is capable of facilitating any-to-any multimodal conversation while achieving performance comparable to specialized models across all modalities, proving that discrete representations can effectively and conveniently unify multiple modalities within a language model. Demos are shown in https://junzhan2000.github.io/AnyGPT.github.io/

Community

cool beans

This is an automated message from the Librarian Bot. I found the following papers similar to this paper.

The following papers were recommended by the Semantic Scholar API

Please give a thumbs up to this comment if you found it helpful!

If you want recommendations for any Paper on Hugging Face checkout this Space

You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend

So this could make an idea that probably won't work but I think might be worth testing. It follows the US constitution. Using this to train a bottom up executive AI. So you have a judge ie a community/humans. Senate top down. Congress bottom up. Executive agent that figures out how to get it done. Research based user focused and results driven. Any thoughts?

This comment has been hidden

A_vibrant_and_eyecatching_digital_illustration_.jpg

c40027ac44de1ea4814b1888e17c2793.jpg

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2402.12226 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2402.12226 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2402.12226 in a Space README.md to link it from this page.

Collections including this paper 29