Datasets:

Modalities:
Image
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
pandas
MMBench_dev / README.md
CopyPaste001's picture
[Feature]: Add dataset card
b6d4c83
|
raw
history blame
5.82 kB
metadata
dataset_info:
  features:
    - name: question
      dtype: string
    - name: hint
      dtype: string
    - name: A
      dtype: string
    - name: B
      dtype: string
    - name: C
      dtype: string
    - name: D
      dtype: string
    - name: label
      dtype:
        class_label:
          names:
            '0': A
            '1': B
            '2': C
            '3': D
    - name: image
      dtype: image
  splits:
    - name: train
      num_bytes: 102942038.498
      num_examples: 4377
  download_size: 99866501
  dataset_size: 102942038.498

Dataset Card for "MMBench_dev"

Dataset Description

Dataset Summary

In recent years, the field has seen a surge in the development of numerous vision-language (VL) models, such as MiniGPT-4 and LLaVA. These models showcase promising performance in tackling previously challenging tasks. However, effectively evaluating these models' performance has become a primary challenge hindering further advancement in large VL models. Traditional benchmarks like VQAv2 and COCO Caption are widely used to provide quantitative evaluations for VL models but suffer from several shortcomings:

Dataset Construction: Dataset Construction: Traditional benchmarks tend to evaluate models based on their performance in various tasks, such as image captioning and visual question answering. Unfortunately, these tasks do not fully capture the fine-grained abilities that a model possesses, potentially impeding future optimization efforts.

Evaluation Metrics: Existing evaluation metrics lack robustness. For example, VQAv2 targets a single word or phrase, while many current VL models generate sentences as outputs. Although these sentences may correctly answer the corresponding questions, the existing evaluation metric would assign a Fail score due to an inability to exactly match the given answer. Moreover, recently proposed subjective evaluation metrics, such as that used in mPLUG-Owl, offer comprehensive evaluation of VL models. However, these metrics struggle to scale smoothly due to the significant amount of human labor required for evaluation. Additionally, these evaluations are highly biased and difficult to reproduce.

To address these limitations, we propose a novel approach by defining a set of fine-grained abilities and collecting relevant questions for each ability. We also introduce innovative evaluation strategies to ensure more robust assessment of model predictions. This new benchmark, called MMBench, boasts the following features:

Data Collection: To date, we have gathered approximately 3000 questions spanning 20 ability dimensions. Each question is a multiple-choice format with a single correct answer.

Evaluation: For a more reliable evaluation, we employ ChatGPT to match a model's prediction with the choices of a question, and then output the corresponding label (A, B, C, D) as the final prediction.

Languages

All of our questions are presented in single-choice question format, with the number of options ranging from 2 to 4. In addition, all these questions, options, and answers are in English.

Dataset Structure

Data Instances

We provide a overview of an instance in MMBench as follows:

{
  'index': 241,
  'question': 'Identify the question that Madelyn and Tucker's experiment can best answer.',
  'hint': 'The passage below describes an experiment. Read the passage and then follow the 
          instructions below.\n\nMadelyn applied a thin layer of wax to the underside of her  
          snowboard and rode the board straight down a hill. Then, she removed the wax and rode 
          the snowboard straight down the hill again. She repeated the rides four more times, 
          alternating whether she rode with a thin layer of wax on the board or not. Her friend 
          Tucker timed each ride. Madelyn and Tucker calculated the average time it took to slide 
          straight down the hill on the snowboard with wax compared to the average time on the 
          snowboard without wax.\nFigure: snowboarding down a hill.'
  'A': 'Does Madelyn's snowboard slide down a hill in less time when it has a thin layer of wax or 
        a thick layer of wax?'
  'B': 'Does Madelyn's snowboard slide down a hill in less time when it has a layer of wax or     
        when it does not have a layer of wax?'
  'image': xxxxxx,
  'category': 'identity_reasoning',
  'l2-category': 'attribute_reasoning',
  'split': 'dev',
  'source': 'scienceqa',
}

Data Fields

  • index: the index of the instance in the dataset.
  • question: the question of the instance.
  • hint (optional): the hint of the instance.
  • A: the first option of the instance.
  • B: the second option of the instance.
  • C (optional): the third option of the instance.
  • D (optional): the fourth option of the instance.
  • image: the raw image of the instance.
  • category: the leaf category of the instance.
  • l2-category: the L-2 category of the instance.
  • split: the split of the instance.
  • source: the source of the instance comes from.

Data Splits

Currently, MMBench contains 2974 instances in total, and is splitted into dev and test splits according to a 4:6 ratio.

Additional Information

Citation Information

@article{MMBench,
    author  = {Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhnag, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin},
    journal = {arXiv:2307.06281},
    title   = {MMBench: Is Your Multi-modal Model an All-around Player?},
    year    = {2023},
}