Can Vision-Language Models Answer Face to Face Questions in the Real-World?
Abstract
AI models have made significant strides in recent years in their ability to describe and answer questions about real-world images. They have also made progress in the ability to converse with users in real-time using audio input. This raises the question: have we reached the point where AI models, connected to a camera and microphone, can converse with users in real-time about scenes and events that are unfolding live in front of the camera? This has been a long-standing goal in AI and is a prerequisite for real-world AI assistants and humanoid robots to interact with humans in everyday situations. In this work, we introduce a new dataset and benchmark, the Qualcomm Interactive Video Dataset (IVD), which allows us to assess the extent to which existing models can support these abilities, and to what degree these capabilities can be instilled through fine-tuning. The dataset is based on a simple question-answering setup, where users ask questions that the system has to answer, in real-time, based on the camera and audio input. We show that existing models fall far behind human performance on this task, and we identify the main sources for the performance gap. However, we also show that for many of the required perceptual skills, fine-tuning on this form of data can significantly reduce this gap.
Community
TL;DR:
Current AI models fail at situated AI, where AI systems must interpret and respond to real-time visual and audio inputs or talk talk to a user in the real world and understand what's happening right now. Towards this open problem, the work presents Interactive Video Dataset (IVD) as a benchmark to evaluate and improve models' capabilities for real-time conversation about live scenes.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- An Egocentric Vision-Language Model based Portable Real-time Smart Assistant (2025)
- WorldSense: Evaluating Real-world Omnimodal Understanding for Multimodal LLMs (2025)
- Exploring Advanced Techniques for Visual Question Answering: A Comprehensive Comparison (2025)
- Rethinking Bottlenecks in Safety Fine-Tuning of Vision Language Models (2025)
- RDMM: Fine-Tuned LLM Models for On-Device Robotic Decision Making with Enhanced Contextual Awareness in Specific Domains (2025)
- On the robustness of multimodal language model towards distractions (2025)
- Open-Ended and Knowledge-Intensive Video Question Answering (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper