Datasets:

Modalities:
Image
Text
Formats:
parquet
ArXiv:
Libraries:
Datasets
pandas
PY007 commited on
Commit
78601bc
1 Parent(s): 28667eb

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +21 -2
README.md CHANGED
@@ -79,6 +79,25 @@ configs:
79
  - split: test
80
  path: data/test-*
81
  ---
82
- # Dataset Card for "MMBench_CN"
 
 
83
 
84
- [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
79
  - split: test
80
  path: data/test-*
81
  ---
82
+ <p align="center" width="100%">
83
+ <img src="https://i.postimg.cc/g0QRgMVv/WX20240228-113337-2x.png" width="100%" height="80%">
84
+ </p>
85
 
86
+ # Large-scale Multi-modality Models Evaluation Suite
87
+
88
+ > Accelerating the development of large-scale multi-modality models (LMMs) with `lmms-eval`
89
+
90
+ 🏠 [Homepage](https://lmms-lab.github.io/) | 📚 [Documentation](docs/README.md) | 🤗 [Huggingface Datasets](https://huggingface.co/lmms-lab)
91
+
92
+ # This Dataset
93
+
94
+ This is a formatted version of the Chinese subset of [MMBench](https://arxiv.org/abs/2307.06281). It is used in our `lmms-eval` pipeline to allow for one-click evaluations of large multi-modality models.
95
+
96
+ ```
97
+ @article{MMBench,
98
+ author = {Yuan Liu, Haodong Duan, Yuanhan Zhang, Bo Li, Songyang Zhang, Wangbo Zhao, Yike Yuan, Jiaqi Wang, Conghui He, Ziwei Liu, Kai Chen, Dahua Lin},
99
+ journal = {arXiv:2307.06281},
100
+ title = {MMBench: Is Your Multi-modal Model an All-around Player?},
101
+ year = {2023},
102
+ }
103
+ ```