--- license: cc-by-2.0 dataset_info: features: - name: sample_id dtype: int32 - name: task_instruction dtype: string - name: task_instance struct: - name: context dtype: string - name: images_path sequence: string - name: choice_list sequence: string - name: combined_1_images sequence: string - name: response dtype: string splits: - name: ActionLocalization_test num_bytes: 291199 num_examples: 200 - name: ActionLocalization_adv num_bytes: 291199 num_examples: 200 - name: ActionPrediction_test num_bytes: 255687 num_examples: 200 - name: ActionPrediction_adv num_bytes: 255687 num_examples: 200 - name: ActionSequence_test num_bytes: 262234 num_examples: 200 - name: ActionSequence_adv num_bytes: 262234 num_examples: 200 - name: ALFRED_test num_bytes: 112715 num_examples: 200 - name: ALFRED_adv num_bytes: 112715 num_examples: 200 - name: CharacterOrder_test num_bytes: 274821 num_examples: 200 - name: CharacterOrder_adv num_bytes: 274821 num_examples: 200 - name: CLEVR_Change_test num_bytes: 114792 num_examples: 200 - name: CLEVR_Change_adv num_bytes: 114792 num_examples: 200 - name: CounterfactualInference_test num_bytes: 129074 num_examples: 200 - name: CounterfactualInference_adv num_bytes: 129074 num_examples: 200 - name: DocVQA_test num_bytes: 76660 num_examples: 200 - name: DocVQA_adv num_bytes: 76660 num_examples: 200 - name: EgocentricNavigation_test num_bytes: 559193 num_examples: 200 - name: EgocentricNavigation_adv num_bytes: 559193 num_examples: 200 - name: GPR1200_test num_bytes: 579624 num_examples: 600 - name: IEdit_test num_bytes: 50907 num_examples: 200 - name: IEdit_adv num_bytes: 50907 num_examples: 200 - name: ImageNeedleInAHaystack_test num_bytes: 303423 num_examples: 320 - name: MMCoQA_test num_bytes: 344623 num_examples: 200 - name: MMCoQA_adv num_bytes: 344623 num_examples: 200 - name: MovingAttribute_test num_bytes: 97299 num_examples: 200 - name: MovingAttribute_adv num_bytes: 97299 num_examples: 200 - name: MovingDirection_test num_bytes: 115832 num_examples: 200 - name: MovingDirection_adv num_bytes: 115832 num_examples: 200 - name: MultiModalQA_test num_bytes: 87978 num_examples: 200 - name: MultiModalQA_adv num_bytes: 87978 num_examples: 200 - name: nuscenes_test num_bytes: 87282 num_examples: 200 - name: nuscenes_adv num_bytes: 87282 num_examples: 200 - name: ObjectExistence_test num_bytes: 94139 num_examples: 200 - name: ObjectExistence_adv num_bytes: 94139 num_examples: 200 - name: ObjectInteraction_test num_bytes: 264032 num_examples: 200 - name: ObjectInteraction_adv num_bytes: 264032 num_examples: 200 - name: ObjectShuffle_test num_bytes: 289186 num_examples: 200 - name: ObjectShuffle_adv num_bytes: 289186 num_examples: 200 - name: OCR_VQA_test num_bytes: 80940 num_examples: 200 - name: OCR_VQA_adv num_bytes: 80940 num_examples: 200 - name: SceneTransition_test num_bytes: 266203 num_examples: 200 - name: SceneTransition_adv num_bytes: 266203 num_examples: 200 - name: SlideVQA_test num_bytes: 89462 num_examples: 200 - name: SlideVQA_adv num_bytes: 89462 num_examples: 200 - name: Spot_the_Diff_test num_bytes: 47823 num_examples: 200 - name: Spot_the_Diff_adv num_bytes: 47823 num_examples: 200 - name: StateChange_test num_bytes: 286783 num_examples: 200 - name: StateChange_adv num_bytes: 286783 num_examples: 200 - name: TextNeedleInAHaystack_test num_bytes: 11140730 num_examples: 320 - name: TQA_test num_bytes: 92861 num_examples: 200 - name: TQA_adv num_bytes: 92861 num_examples: 200 - name: WebQA_test num_bytes: 202682 num_examples: 200 - name: WebQA_adv num_bytes: 202682 num_examples: 200 - name: WikiVQA_test num_bytes: 2557847 num_examples: 200 - name: WikiVQA_adv num_bytes: 2557847 num_examples: 200 download_size: 12035444 dataset_size: 26288285 configs: - config_name: default data_files: - split: ActionLocalization_test path: preview/ActionLocalization_test-* - split: ActionLocalization_adv path: preview/ActionLocalization_adv-* - split: ActionPrediction_test path: preview/ActionPrediction_test-* - split: ActionPrediction_adv path: preview/ActionPrediction_adv-* - split: ActionSequence_test path: preview/ActionSequence_test-* - split: ActionSequence_adv path: preview/ActionSequence_adv-* - split: ALFRED_test path: preview/ALFRED_test-* - split: ALFRED_adv path: preview/ALFRED_adv-* - split: CharacterOrder_test path: preview/CharacterOrder_test-* - split: CharacterOrder_adv path: preview/CharacterOrder_adv-* - split: CLEVR_Change_test path: preview/CLEVR_Change_test-* - split: CLEVR_Change_adv path: preview/CLEVR_Change_adv-* - split: CounterfactualInference_test path: preview/CounterfactualInference_test-* - split: CounterfactualInference_adv path: preview/CounterfactualInference_adv-* - split: DocVQA_test path: preview/DocVQA_test-* - split: DocVQA_adv path: preview/DocVQA_adv-* - split: EgocentricNavigation_test path: preview/EgocentricNavigation_test-* - split: EgocentricNavigation_adv path: preview/EgocentricNavigation_adv-* - split: GPR1200_test path: preview/GPR1200_test-* - split: IEdit_test path: preview/IEdit_test-* - split: IEdit_adv path: preview/IEdit_adv-* - split: ImageNeedleInAHaystack_test path: preview/ImageNeedleInAHaystack_test-* - split: MMCoQA_test path: preview/MMCoQA_test-* - split: MMCoQA_adv path: preview/MMCoQA_adv-* - split: MovingAttribute_test path: preview/MovingAttribute_test-* - split: MovingAttribute_adv path: preview/MovingAttribute_adv-* - split: MovingDirection_test path: preview/MovingDirection_test-* - split: MovingDirection_adv path: preview/MovingDirection_adv-* - split: MultiModalQA_test path: preview/MultiModalQA_test-* - split: MultiModalQA_adv path: preview/MultiModalQA_adv-* - split: nuscenes_test path: preview/nuscenes_test-* - split: nuscenes_adv path: preview/nuscenes_adv-* - split: ObjectExistence_test path: preview/ObjectExistence_test-* - split: ObjectExistence_adv path: preview/ObjectExistence_adv-* - split: ObjectInteraction_test path: preview/ObjectInteraction_test-* - split: ObjectInteraction_adv path: preview/ObjectInteraction_adv-* - split: ObjectShuffle_test path: preview/ObjectShuffle_test-* - split: ObjectShuffle_adv path: preview/ObjectShuffle_adv-* - split: OCR_VQA_test path: preview/OCR_VQA_test-* - split: OCR_VQA_adv path: preview/OCR_VQA_adv-* - split: SceneTransition_test path: preview/SceneTransition_test-* - split: SceneTransition_adv path: preview/SceneTransition_adv-* - split: SlideVQA_test path: preview/SlideVQA_test-* - split: SlideVQA_adv path: preview/SlideVQA_adv-* - split: Spot_the_Diff_test path: preview/Spot_the_Diff_test-* - split: Spot_the_Diff_adv path: preview/Spot_the_Diff_adv-* - split: StateChange_test path: preview/StateChange_test-* - split: StateChange_adv path: preview/StateChange_adv-* - split: TextNeedleInAHaystack_test path: preview/TextNeedleInAHaystack_test-* - split: TQA_test path: preview/TQA_test-* - split: TQA_adv path: preview/TQA_adv-* - split: WebQA_test path: preview/WebQA_test-* - split: WebQA_adv path: preview/WebQA_adv-* - split: WikiVQA_test path: preview/WikiVQA_test-* - split: WikiVQA_adv path: preview/WikiVQA_adv-* task_categories: - visual-question-answering - question-answering - text-generation - image-to-text - video-classification language: - en tags: - Long-context - MLLM - VLM - LLM - Benchmark pretty_name: MileBench size_categories: - 1K To construct our evaluation sets, we gather 6,440 multimodal long-context samples from 21 pre-existing or self-constructed datasets, with an average of 15.2 images and 422.3 words each, as depicted in the figure, and we categorize them into their respective subsets.
stat2stat1
## How to use? Please download MileBench_part*.tar.gz and unzip them using the following command. ```bash for file in MileBench_part*.tar.gz do tar -xzvf "$file" done ``` Then please refer to [Code for MileBench](https://github.com/MileBench/MileBench?tab=readme-ov-file#-dataset-preparation) to evaluate. ## Links - **Homepage:** [MileBench Homepage](https://milebench.github.io/) - **Repository:** [MileBench GitHub](https://github.com/MileBench/MileBench) - **Paper:** [Arxiv](https://arxiv.org/abs/2404.18532) - **Point of Contact:** [Dingjie Song](mailto:bbsngg@outlook.com) ## Citation If you find this project useful in your research, please consider citing: ```BibTeX @article{song2024milebench, title={MileBench: Benchmarking MLLMs in Long Context}, author={Song, Dingjie and Chen, Shunian and Chen, Guiming Hardy and Yu, Fei and Wan, Xiang and Wang, Benyou}, journal={arXiv preprint arXiv:2404.18532}, year={2024} } ```