Edit model card

DeciDiffusion 2.0

DeciDiffusion 2.0 is a 732 million parameter text-to-image latent diffusion model, generated with the help of AutoNAC, Deci's proprietary Neural Architecture Search technology. Advanced training techniques were used to speed up training, improve training performance, and achieve better inference quality.

Model Details

  • Developed by: Deci
  • Model type: Diffusion-based text-to-image generation model
  • Language(s) (NLP): English
  • License: The model is released under the CreativeML Open RAIL++-M license.

Model Resources

Model Architecture

DeciDiffusion 2.0, a state-of-the-art diffusion-based text-to-image generation model, builds upon the core architecture of Stable Diffusion. It incorporates key elements like the Variational Autoencoder (VAE) and the pre-trained Text Encoder CLIP. A standout feature of DeciDiffusion is its U-Net component, which is optimized for performance on cost-effective hardware. DeciDiffusion’s AutoNAC-generated U-Net-NAS features 525 million parameters as opposed to the 860 in Stable Diffusion 1.5’s U-Net. This optimized design significantly enhances processing speed, making DeciDiffusion a highly efficient and effective solution in the realm of text-to-image generation.

Training Details

Training Procedure

The model was trained in 4 phases:

  • Phase 1: Trained from scratch 1.28 million steps at resolution 256x256.
  • Phase 2: Trained from 870k steps at resolution 512x512 on the same dataset to learn more fine-detailed information.
  • Phase 3: Trained 65k steps with EMA, another learning rate scheduler, and more "qualitative" data.
  • Phase 4: Fine-tuning on a 2M sample dataset.

Training Techniques

DeciDiffusion 2.0 marks a significant advancement over previous latent diffusion models, particularly in terms of sample efficiency. This means it can produce high-quality images with fewer diffusion timesteps during the inference process. To attain such efficiency, Deci has refined the DPM++ scheduler, effectively cutting down the number of steps needed to generate a quality image from 16 to just 10.

Additionally, the following training techniques were used to improve the model's sample efficiency:

Additional Details

Phase 1

  • Hardware: 6 x 8 x H100 (80GB)
  • Optimizer: LAMB
  • Batch: 18432
  • Learning rate: 5e-03

Phases 2-4

  • Hardware: 8 x 8 x H100 (80GB)
  • Optimizer: LAMB
  • Batch: 7168
  • Learning rate: 5e-03

Runtime Benchmarks

The following tables provide an image latency comparison between DeciDiffusion 2.0 and Stable Diffusion v1.5.

DeciDiffusion 2.0 vs. Stable Diffusion v1.5 at FP16 precision

Implementation + Iterations DeciDiffusion 2.0 on AI 100 (seconds/image) Stable Diffusion v1.5 on AI 100 (seconds/image)
Compiled 16 Iterations 1.335 2.478
Compiled 10 Iterations 0.971 1.684

How to Use

Note: You must use diffusers v0.21.4 to run the model successfully.

# pip install diffusers==0.21.4 transformers torch

from diffusers import StableDiffusionPipeline
import torch

device = 'cuda' if torch.cuda.is_available() else 'cpu'

checkpoint = "Deci/DeciDiffusion-v2-0"
pipeline = StableDiffusionPipeline.from_pretrained(checkpoint, custom_pipeline=checkpoint, torch_dtype=torch.float16)
pipeline.unet = pipeline.unet.from_pretrained(checkpoint, subfolder='flexible_unet', torch_dtype=torch.float16)

pipeline = pipeline.to(device)

img = pipeline(prompt=['A photo of an astronaut riding a horse on Mars']).images[0]


Misuse, Malicious Use, and Out-of-Scope Use

The model must not be employed to deliberately produce or spread images that foster hostile or unwelcoming settings for individuals. This encompasses generating visuals that might be predictably upsetting, distressing, or inappropriate, as well as content that perpetuates existing or historical biases.

Out-of-Scope Use

The model isn't designed to produce accurate or truthful depictions of people or events. Thus, using it for such purposes exceeds its intended capabilities.

Misuse and Malicious Use

Misusing the model to produce content that harms or maligns individuals is strictly discouraged. Such misuses include, but aren't limited to:

  • Creating offensive, degrading, or damaging portrayals of individuals, their cultures, religions, or surroundings.
  • Intentionally promoting or propagating discriminatory content or harmful stereotypes.Deliberately endorsing or disseminating prejudiced content or harmful stereotypes.
  • Deliberately endorsing or disseminating prejudiced content or harmful stereotypes.
  • Posing as someone else without their agreement.
  • Generating explicit content without the knowledge or agreement of potential viewers.
  • Distributing copyrighted or licensed content against its usage terms.
  • Sharing modified versions of copyrighted or licensed content in breach of its usage guidelines.

Limitations and Bias


The model has certain limitations and may not function optimally in the following scenarios:

  • It doesn't produce completely photorealistic images.
  • Rendering legible text is beyond its capability.
  • Complex compositions, like visualizing “A green sphere to the left of a blue square”, are challenging for the model.
  • Generation of faces and human figures may be imprecise.
  • It is primarily optimized for English captions and might not be as effective with other languages.
  • The autoencoding component of the model is lossy.


The remarkable abilities of image-generation models can unintentionally amplify societal biases. DeciDiffusion was trained with a focus on English descriptions. Consequently, non-English communities and cultures might be underrepresented, leading to a bias towards white and western norms. Outputs from non-English prompts are notably less accurate. Given these biases, users should approach DeciDiffusion with discretion, regardless of input.

How to Cite

Please cite this model using this format.

title = {DeciDiffusion 2.0},
author = {DeciAI Research Team},
year = {2024}
Downloads last month

Spaces using Deci/DeciDiffusion-v2-0 2

Collection including Deci/DeciDiffusion-v2-0