Text-to-Image
Diffusers
ONNX
Safetensors
English
StableDiffusionXLPipeline
common-canvas
stable-diffusion
sdxl
Inference Endpoints
Skylion007 commited on
Commit
a844463
1 Parent(s): cb710c9

Add Model Card

Browse files
Files changed (1) hide show
  1. README.md +55 -2
README.md CHANGED
@@ -1,6 +1,59 @@
1
  ---
2
- license: openrail++
3
  tags:
4
- - text-to-image
5
  - stable-diffusion
 
 
 
 
 
 
6
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ license: cc-by-nc-sa-4.0
3
  tags:
4
+ - common-canvas
5
  - stable-diffusion
6
+ - sdxl
7
+ datasets:
8
+ - common-canvas/commoncatalog-cc-by-sa
9
+ - common-canvas/commoncatalog-cc-by
10
+ language:
11
+ - en
12
  ---
13
+ # CommonCanvas-XLC
14
+
15
+ ## Summary
16
+ CommonCanvas is a family of latent diffusion models capable of generating images from a given text prompt. The architecture is based off of Stable Diffusion XL. Different CommonCanvas models are trained exclusively on subsets of the CommonCatalog Dataset (See Data Card), a large dataset of Creative Commons licensed images with synthetic captions produced using a pre-trained BLIP-2 captioning model.
17
+
18
+ Input: CommonCatalog Text Captions
19
+ Output: CommonCatalog Images
20
+ Architecture: Stable Diffusion XL
21
+ Version Number: 0.1
22
+
23
+ The goal of this purpose is to produce a model that is competitive with Stable Diffusion XL, but to do so using an easily accessible dataset of known provenance. Doing so makes replicating the model significantly easier and provides proper attribution to all the creative commons work used to train the model. The exact training recipe of the model can be found in the paper hosted at this link. https://arxiv.org/abs/2310.16825
24
+
25
+ ## Performance Limitations
26
+
27
+ CommonCanvas under-performs in several categories, including faces, general photography, and paintings (see paper, Figure 8). These datasets all originated from the Conceptual Captions dataset, which relies on web-scraped data. These web-sourced captions, while abundant, may not always align with human-generated language nuances. Transitioning to synthetic captions introduces certain performance challenges, however, the drop in performance is not as dramatic as one might assume.
28
+
29
+ ## Training Dataset Limitations
30
+ The model is trained on 10 year old YFCC data and may not have modern concepts or recent events in its training corpus. Performance on this model will be worse on certain proper nouns or specific celebrities, but this is a feature not a bug. The model may not generate known artwork, individual celebrities, or specific locations due to the autogenerated nature of the caption data.
31
+
32
+
33
+ Note: The non-commercial variants of this model are explicitly not intended to be use
34
+ * It is trained on data derived from the Flickr100M dataset. The information is dated and known to have a bias towards internet connected Western countries. Some areas such as the global south lack representation.
35
+
36
+ ## Associated Risks
37
+ * Text in images produced by the model will likely be difficult to read.
38
+ * The model struggles with more complex tasks that require compositional understanding
39
+ * It may not accurately generate faces or representations of specific people.
40
+ * The model primarily learned from English descriptions and may not perform as effectively in other languages.
41
+ * The autoencoder aspect of the model introduces some information loss.
42
+ * It may be possible to guide the model to generate objectionable content, i.e. nudity or other NSFW material.
43
+
44
+ ## Intended Uses
45
+ * Using the model for generative AI research
46
+ * Safe deployment of models which have the potential to generate harmful content.
47
+ * Probing and understanding the limitations and biases of generative models.
48
+ * Generation of artworks and use in design and other artistic processes.
49
+ * Applications in educational or creative tools.
50
+ * Research on generative models.
51
+
52
+ ## Usage
53
+ We recommend using the MosaicML Diffusion Repo to finetune / train the model: https://github.com/mosaicml/diffusion . Example finetuning code coming soon. See the Gradio Demo [Here](https://github.com/mosaicml/diffusion/blob/main/scripts/gradio_demo.py)
54
+
55
+ ## Evaluation/Validation
56
+ We validated the model against Stability AI’s SD2 model and compared human user study
57
+
58
+ ## Acknowledgements
59
+ We thank @multimodalart, @Wauplin, and @lhoestq at Hugging Face for helping us host the dataset, and model weights.