Teaching Metric Distance to Autoregressive Multimodal Foundational Models
Abstract
As large language models expand beyond natural language to domains such as mathematics, multimodal understanding, and embodied agents, tokens increasingly reflect metric relationships rather than purely linguistic meaning. We introduce DIST2Loss, a distance-aware framework designed to train autoregressive discrete models by leveraging predefined distance relationships among output tokens. At its core, DIST2Loss transforms continuous exponential family distributions derived from inherent distance metrics into discrete, categorical optimization targets compatible with the models' architectures. This approach enables the models to learn and preserve meaningful distance relationships during token generation while maintaining compatibility with existing architectures. Empirical evaluations show consistent performance gains in diverse multimodal applications, including visual grounding, robotic manipulation, generative reward modeling, and image generation using vector-quantized features. These improvements are pronounced in cases of limited training data, highlighting DIST2Loss's effectiveness in resource-constrained settings.
Community
TL;DR: Our study investigates techniques for embedding output-space metric priors directly into the loss functions of multimodal autoregressive foundational models. This integration accelerates learning in tasks where metrics are crucial—for example, coordinate prediction in object detection, action determination in object manipulation, and Likert scale scoring in reward modeling.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- MAGNET: Augmenting Generative Decoders with Representation Learning and Infilling Capabilities (2025)
- Probabilistic Subspace Manifolds for Contextual Inference in Large Language Models (2025)
- VARGPT: Unified Understanding and Generation in a Visual Autoregressive Multimodal Large Language Model (2025)
- Contextual Reinforcement in Multimodal Token Compression for Large Language Models (2025)
- Neural Contextual Reinforcement Framework for Logical Structure Language Generation (2025)
- IPO: Your Language Model is Secretly a Preference Classifier (2025)
- Scaling LLM Pre-training with Vocabulary Curriculum (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper