Datasets:

Modalities:
Image
ArXiv:
License:
Dataset Preview
Viewer
The full dataset viewer is not available (click to read why). Only showing a preview of the rows.
(ReadTimeoutError("HTTPSConnectionPool(host='huggingface.co', port=443): Read timed out. (read timeout=10)"), '(Request ID: 75af9d77-2df3-48bd-95ff-a6c8e2b3fe97)')
Error code:   UnexpectedError

Need help to make the dataset viewer work? Open a discussion for direct support.

image
image
End of preview.

What is the Visual Cognition Gap between Humans and Multimodal LLMs?

Description:

VCog-Bench is a publicly available zero-shot abstract visual reasoning (AVR) benchmark designed to evaluate Multimodal Large Language Models (MLLMs). This benchmark integrates two well-known AVR datasets from the AI community and includes a newly proposed MaRs-VQA dataset. The findings in VCog-Bench show that current state-of-the-art MLLMs and Vision-Language Models (VLMs), such as GPT-4o and LLaVA-1.6, InternVL demonstrate some basic understanding of AVR tasks. However, these models still face challenges with complex matrix reasoning tasks. This highlights the need for further exploration and development in this area. By providing a robust benchmark, we aim to encourage further innovation and progress in the field of zero-shot abstract visual reasoning.

Benchmark Dataset Structure:

----vcog-bench
    |----cvr
    |    |----case_name1
    |    |    |----answer
    |    |    |    |----image
    |    |    |    |    |----x.png
    |    |    |----choice
    |    |    |    |----image
    |    |    |    |    |----sub_image_0.png
    |    |    |    |    |----sub_image_1.png
    |    |    |    |    |----sub_image_2.png
    |    |    |    |    |----sub_image_3.png
    |    |----case_name2
    |    |----case_name3
    |    |----case_name4
    |    |----......
    |----raven
    |    |----case_name1
    |    |    |----answer
    |    |    |    |----image
    |    |    |    |    |----x.jpeg
    |    |    |----choice
    |    |    |    |----image
    |    |    |    |    |----0.jpeg
    |    |    |    |    |----1.jpeg
    |    |    |    |    |----2.jpeg
    |    |    |    |    |----3.jpeg
    |    |    |    |    |----4.jpeg
    |    |    |    |    |----5.jpeg
    |    |    |    |    |----6.jpeg
    |    |    |    |    |----7.jpeg
    |    |    |    |----text
    |    |    |    |    |----annotation.json
    |    |    |----question
    |    |    |    |----image
    |    |    |    |    |----question.jpeg
    |    |----case_name2
    |    |----case_name3
    |    |----case_name4
    |    |----......
    |----marsvqa
    |    |----case_name1
    |    |    |----answer
    |    |    |    |----image
    |    |    |    |    |----xxx.jpeg
    |    |    |----choice
    |    |    |    |----image
    |    |    |    |    |----xxx.jpeg
    |    |    |    |    |----xxx.jpeg
    |    |    |    |    |----xxx.jpeg
    |    |    |    |    |----xxx.jpeg
    |    |    |    |----text
    |    |    |    |    |----annotation.json
    |    |    |----choiceX
    |    |    |    |----image
    |    |    |    |    |----xxx.jpeg
    |    |    |    |    |----xxx.jpeg
    |    |    |    |    |----xxx.jpeg
    |    |    |    |    |----xxx.jpeg
    |    |    |----question
    |    |    |    |----image
    |    |    |    |    |----xxx.jpeg
    |    |----case_name2
    |    |----case_name3
    |    |----case_name4
    |    |----......

Dataset Details

Content Types: VQA pairs with multiple images input
Volume: 560 VQA pairs (RAVEN), 480 VQA pairs (MaRs-VQA), 309 VQA pairs (CVR)
Source of Data: RAVEN dataset, MaRs-IB, CVR dataset
Data Collection Method: See the paper.

Reference

@misc{cao2024visualcognitiongaphumans,
      title={What is the Visual Cognition Gap between Humans and Multimodal LLMs?}, 
      author={Xu Cao and Bolin Lai and Wenqian Ye and Yunsheng Ma and Joerg Heintz and Jintai Chen and Jianguo Cao and James M. Rehg},
      year={2024},
      eprint={2406.10424},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2406.10424}, 
}
Downloads last month
0