Files changed (1) hide show
  1. README.md +94 -40
README.md CHANGED
@@ -1,40 +1,94 @@
1
- # Model Card: DALL·E Mini
2
-
3
- This model is a reproduction of OpenAI’s DALL·E. Please see [this link](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini-Generate-images-from-any-text-prompt--VmlldzoyMDE4NDAy) for project-specific details. Below, we include the original DALL·E model card available on [the OpenAI github](https://github.com/openai/DALL-E/edit/master/model_card.md).
4
-
5
- ## Model Details
6
-
7
- The dVAE was developed by researchers at OpenAI to reduce the memory footprint of the transformer trained on the
8
- text-to-image generation task. The details involved in training the dVAE are described in [the paper][dalle_paper]. This
9
- model card describes the first version of the model, released in February 2021. The model consists of a convolutional
10
- encoder and decoder whose architectures are described [here](dall_e/encoder.py) and [here](dall_e/decoder.py), respectively.
11
- For questions or comments about the models or the code release, please file a Github issue.
12
-
13
- ## Model Use
14
-
15
- ### Intended Use
16
-
17
- The model is intended for others to use for training their own generative models.
18
-
19
- ### Out-of-Scope Use Cases
20
-
21
- This model is inappropriate for high-fidelity image processing applications. We also do not recommend its use as a
22
- general-purpose image compressor.
23
-
24
- ## Training Data
25
-
26
- The model was trained on publicly available text-image pairs collected from the internet. This data consists partly of
27
- [Conceptual Captions][cc] and a filtered subset of [YFCC100M][yfcc100m]. We used a subset of the filters described in
28
- [Sharma et al.][cc_paper] to construct this dataset; further details are described in [our paper][dalle_paper]. We will
29
- not be releasing the dataset.
30
-
31
- ## Performance and Limitations
32
-
33
- The heavy compression from the encoding process results in a noticeable loss of detail in the reconstructed images. This
34
- renders it inappropriate for applications that require fine-grained details of the image to be preserved.
35
-
36
- [dalle_paper]: https://arxiv.org/abs/2102.12092
37
- [cc]: https://ai.google.com/research/ConceptualCaptions
38
- [cc_paper]: https://www.aclweb.org/anthology/P18-1238/
39
- [yfcc100m]: http://projects.dfki.uni-kl.de/yfcc100m/
40
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ int={2102.12092},
2
+ archivePrefix={arXiv},
3
+ primaryClass={cs.CV}
4
+ }
5
+ @misc{
6
+ title={Learning Transferable Visual Models From Natural Language Supervision},
7
+ author={Alec Radford and Jong Wook Kim and Chris Hallacy and Aditya Ramesh and Gabriel Goh and Sandhini Agarwal and Girish Sastry and Amanda Askell and Pamela Mishkin and Jack Clark and Gretchen Krueger and Ilya Sutskever},
8
+ year={2021},
9
+ eprint={2103.00020},
10
+ archivePrefix={arXiv},
11
+ primaryClass={cs.CV}
12
+ }
13
+ @misc{
14
+ title={Taming Transformers for High-Resolution Image Synthesis},
15
+ author={Patrick Esser and Robin Rombach and Björn Ommer},
16
+ year={2021},
17
+ eprint={2012.09841},
18
+ archivePrefix={arXiv},
19
+ primaryClass={cs.CV}
20
+ }
21
+ @misc{
22
+ title={BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension},
23
+ author={Mike Lewis and Yinhan Liu and Naman Goyal and Marjan Ghazvininejad and Abdelrahman Mohamed and Omer Levy and Ves Stoyanov and Luke Zettlemoyer},
24
+ year={2019},
25
+ eprint={1910.13461},
26
+ archivePrefix={arXiv},
27
+ primaryClass={cs.CL}
28
+ }
29
+ @misc{
30
+ title={Scalable Second Order Optimization for Deep Learning},
31
+ author={Rohan Anil and Vineet Gupta and Tomer Koren and Kevin Regan and Yoram Singer},
32
+ year={2021},
33
+ eprint={2002.09018},
34
+ archivePrefix={arXiv},
35
+ primaryClass={cs.LG}
36
+ }
37
+ @misc{
38
+ title={GLU Variants Improve Transformer},
39
+ author={Noam Shazeer},
40
+ year={2020},
41
+ url={https://arxiv.org/abs/2002.05202}
42
+ }
43
+ @misc{
44
+ title={DeepNet: Scaling transformers to 1,000 layers},
45
+ author={Wang, Hongyu and Ma, Shuming and Dong, Li and Huang, Shaohan and Zhang, Dongdong and Wei, Furu},
46
+ year={2022},
47
+ eprint={2203.00555}
48
+ archivePrefix={arXiv},
49
+ primaryClass={cs.LG}
50
+ }
51
+ @misc{
52
+ title={NormFormer: Improved Transformer Pretraining with Extra Normalization},
53
+ author={Sam Shleifer and Jason Weston and Myle Ott},
54
+ year={2021},
55
+ eprint={2110.09456},
56
+ archivePrefix={arXiv},
57
+ primaryClass={cs.CL}
58
+ }
59
+ @inproceedings{
60
+ title={Swin Transformer V2: Scaling Up Capacity and Resolution},
61
+ author={Ze Liu and Han Hu and Yutong Lin and Zhuliang Yao and Zhenda Xie and Yixuan Wei and Jia Ning and Yue Cao and Zheng Zhang and Li Dong and Furu Wei and Baining Guo},
62
+ booktitle={International Conference on Computer Vision and Pattern Recognition (CVPR)},
63
+ year={2022}
64
+ }
65
+ @misc{
66
+ title = {CogView: Mastering Text-to-Image Generation via Transformers},
67
+ author = {Ming Ding and Zhuoyi Yang and Wenyi Hong and Wendi Zheng and Chang Zhou and Da Yin and Junyang Lin and Xu Zou and Zhou Shao and Hongxia Yang and Jie Tang},
68
+ year = {2021},
69
+ eprint = {2105.13290},
70
+ archivePrefix = {arXiv},
71
+ primaryClass = {cs.CV}
72
+ }
73
+ @misc{
74
+ title = {Root Mean Square Layer Normalization},
75
+ author = {Biao Zhang and Rico Sennrich},
76
+ year = {2019},
77
+ eprint = {1910.07467},
78
+ archivePrefix = {arXiv},
79
+ primaryClass = {cs.LG}
80
+ }
81
+ @misc{
82
+ title = {Sinkformers: Transformers with Doubly Stochastic Attention},
83
+ url = {https://arxiv.org/abs/2110.11773},
84
+ author = {Sander, Michael E. and Ablin, Pierre and Blondel, Mathieu and Peyré, Gabriel},
85
+ publisher = {arXiv},
86
+ year = {2021},
87
+ }
88
+ @misc{
89
+ title = {Smooth activations and reproducibility in deep networks},
90
+ url = {https://arxiv.org/abs/2010.09931},
91
+ author = {Shamir, Gil I. and Lin, Dong and Coviello, Lorenzo},
92
+ publisher = {arXiv},
93
+ year = {2020},
94
+ }