Datasets:

Modalities:
Image
Text
Languages:
English
Size:
< 1K
ArXiv:
Libraries:
Datasets
License:
mbrack commited on
Commit
50b68b0
1 Parent(s): e46d820

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -0
README.md CHANGED
@@ -1,3 +1,27 @@
1
  ---
2
  license: mit
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
+ task_categories:
4
+ - text-to-image
5
+ - image-to-image
6
+ language:
7
+ - en
8
+ size_categories:
9
+ - n<1K
10
  ---
11
+ # Multimodal Concept Conjunction 250
12
+
13
+ In our paper [MultiFusion: Fusing Pre-Trained Models for
14
+ Multi-Lingual, Multi-Modal Image Generation](https://arxiv.org/abs/2305.15296) we propose the MCC-250 benchmark to evaluate generative image composition capablities for multimodal inputs.
15
+ MCC-250 is built on a subset of [CC-500](https://arxiv.org/abs/2212.05032) which contains 500 text-only prompts of the pattern "a red apple and a yellow banana", textually
16
+ describing two objects with respective attributes.
17
+
18
+ With MCC-250, we provide a set of reference images for each object and attribute combination, enabling multimodal applications.
19
+
20
+ ## Attribution
21
+
22
+ All images where source from these four stock imagery providers:
23
+ - [Pixabay](https://pixabay.com/)
24
+ - [Unsplash](https://unsplash.com/)
25
+ - [Pexels](https://www.pexels.com/)
26
+ - [Freepik](https://www.freepik.com/)
27
+