Datasets:

Modalities:
Text
Video
Formats:
parquet
Languages:
English
ArXiv:
Libraries:
Datasets
pandas
License:

You need to agree to share your contact information to access this dataset

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this dataset content.

arXiv Website GitHub Code

Visual Spatial Intelligence Benchmark (VSI-Bench)

This repository contains the visual spatial intelligence benchmark (VSI-Bench), introduced in Thinking in Space: How Multimodal Large Language Models See, Remember and Recall Spaces.

Files

The test-00000-of-00001.parquet file contains the complete dataset annotations and pre-loaded images, ready for processing with HF Datasets. It can be loaded using the following code:

from datasets import load_dataset
vsi_bench = load_dataset("nyu-visionx/VSI-Bench")

Additionally, we provide the videos in *.zip.

Dataset Description

VSI-Bench quantitatively evaluates the visual-spatial intelligence of MLLMs from egocentric video. VSI-Bench comprises over 5,000 question-answer pairs derived from 288 real videos. These videos are sourced from the validation sets of the public indoor 3D scene reconstruction datasets ScanNet, ScanNet++, and ARKitScenes, and represent diverse environments -- including residential spaces, professional settings (e.g., offices, labs), and industrial spaces (e.g., factories) and multiple geographic regions. By repurposing these existing 3D reconstruction and understanding datasets, VSI-Bench benefits from accurate object-level annotations, which are used in question generation and could support future studies exploring the connection between MLLMs and 3D reconstruction.

The dataset contains the following fields:

Field Name Description
idx Global index of the entry in the dataset
dataset Video source: scannet, arkitscenes or scannetpp
scene_name Scene (video) name for each question-answer pair
question_type The type of task for question
question Question asked about the video
options Choices for the question (only for multiple choice questions)
ground_truth Ground truth answer for the question

Evaluation

VSI-Bench evaluates performance using two metrics: for multiple-choice questions, we use Accuracy, calculated based on exact matches. For numerical-answer questions, we introduce a new metric, MRA (Mean Relative Accuracy), to assess how closely model predictions align with ground truth values.

We provide an out-of-the-box evaluation of VSI-Bench in our GitHub repository, including the metrics implementation used in our framework. For further detailes, users can refer to our paper and GitHub repository.

Citation

@article{yang2024think,
    title={{Thinking in Space: How Multimodal Large Language Models See, Remember and Recall Spaces}},
    author={Yang, Jihan and Yang, Shusheng and Gupta, Anjali and Han, Rilyn and Fei-Fei, Li and Xie, Saining},
    year={2024},
    journal={arXiv preprint arXiv:2412.14171},
}
Downloads last month
1