multimodalart HF staff commited on
Commit
d20090a
1 Parent(s): c228894

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -7,14 +7,14 @@ tags:
7
 
8
  # Stable Diffusion v2 Model Card
9
  This model card focuses on the model associated with the Stable Diffusion v2 model, available [here](https://github.com/Stability-AI/stablediffusion).
10
- The model is resumed from `512-base-ema.ckpt` and trained for 150k steps using a [v-objective](https://arxiv.org/abs/2202.00512) on the same dataset. Resumed for another 140k steps on `768x768` images.
11
 
12
 
13
  ## Model Details
14
  - **Developed by:** Robin Rombach, Patrick Esser
15
  - **Model type:** Diffusion-based text-to-image generation model
16
  - **Language(s):** English
17
- - **License:** CreativeML Open RAIL++-M License
18
  - **Model Description:** This is a model that can be used to generate and modify images based on text prompts. It is a [Latent Diffusion Model](https://arxiv.org/abs/2112.10752) that uses a fixed, pretrained text encoder ([OpenCLIP-ViT/H](https://github.com/mlfoundations/open_clip)).
19
  - **Resources for more information:** [GitHub Repository](https://github.com/Stability-AI/).
20
  - **Cite as:**
 
7
 
8
  # Stable Diffusion v2 Model Card
9
  This model card focuses on the model associated with the Stable Diffusion v2 model, available [here](https://github.com/Stability-AI/stablediffusion).
10
+ The model is resumed from [stable-diffusion-2-base](https://huggingface.co/stabilityai/stable-diffusion-2-base) (`512-base-ema.ckpt`) and trained for 150k steps using a [v-objective](https://arxiv.org/abs/2202.00512) on the same dataset. Resumed for another 140k steps on `768x768` images.
11
 
12
 
13
  ## Model Details
14
  - **Developed by:** Robin Rombach, Patrick Esser
15
  - **Model type:** Diffusion-based text-to-image generation model
16
  - **Language(s):** English
17
+ - **License:** [CreativeML Open RAIL++-M License](https://huggingface.co/stabilityai/stable-diffusion-2/blob/main/LICENSE-MODEL)
18
  - **Model Description:** This is a model that can be used to generate and modify images based on text prompts. It is a [Latent Diffusion Model](https://arxiv.org/abs/2112.10752) that uses a fixed, pretrained text encoder ([OpenCLIP-ViT/H](https://github.com/mlfoundations/open_clip)).
19
  - **Resources for more information:** [GitHub Repository](https://github.com/Stability-AI/).
20
  - **Cite as:**