Diffusion models have significantly advanced the fields of image, audio, and video generation, but they depend on an iterative sampling process that causes slow generation. To overcome this limitation, we propose consistency models, a new family of models that generate high quality samples by directly mapping noise to data. They support fast one-step generation by design, while still allowing multistep sampling to trade compute for sample quality. They also support zero-shot data editing, such as image inpainting, colorization, and super-resolution, without requiring explicit training on these tasks. Consistency models can be trained either by distilling pre-trained diffusion models, or as standalone generative models altogether. Through extensive experiments, we demonstrate that they outperform existing distillation techniques for diffusion models in one- and few-step sampling, achieving the new state-of-the-art FID of 3.55 on CIFAR-10 and 6.20 on ImageNet 64 x 64 for one-step generation. When trained in isolation, consistency models become a new family of generative models that can outperform existing one-step, non-adversarial generative models on standard benchmarks such as CIFAR-10, ImageNet 64 x 64 and LSUN 256 x 256.
Intuitively, a consistency model can be thought of as a model which, when evaluated on a noisy image and timestep, returns an output image sample similar to that which would be returned by running a sampling algorithm on a diffusion model. Consistency models can be parameterized by any neural network whose input has the same dimensionality as its output, such as a U-Net.
More precisely, given a teacher diffusion model and fixed sampler, we can train ("distill") a consistency model such that when it is given a noisy image and its corresponding timestep, the output sample of the consistency model will be close to the output that would result by using the sampler on the diffusion model to produce a sample, starting at the same noisy image and timestep. The authors call this procedure "consistency distillation (CD)". Consistency models can also be trained from scratch to generate clean images from a noisy image and timestep, which the authors call "consistency training (CT)".
This model is a
diffusers-compatible version of the cd_imagenet64_lpips.pt checkpont from the original code and model release.
This model was distilled (via consistency distillation (CD)) from an EDM model trained on the ImageNet 64x64 dataset, using LPIPS as the measure of closeness.
See the original model card for more information.
The original PyTorch model checkpoint can be downloaded from the original code and model release.
diffusers pipeline for the
cd-imagenet64-lpips model can be downloaded as follows:
from diffusers import ConsistencyModelPipeline pipe = ConsistencyModelPipeline.from_pretrained("dg845/diffusers-cd_imagenet64_lpips")
The original model checkpoint can be used with the original consistency models codebase.
Here is an example of using the
cd_imagenet64_lpips checkpoint with
import torch from diffusers import ConsistencyModelPipeline device = "cuda" # Load the cd_imagenet64_lpips checkpoint. model_id_or_path = "dg845/diffusers-cd_imagenet64_lpips" pipe = ConsistencyModelPipeline.from_pretrained(model_id_or_path, torch_dtype=torch.float16) pipe.to(device) # Onestep Sampling image = pipe(num_inference_steps=1).images image.save("cd_imagenet64_lpips_onestep_sample.png") # Onestep sampling, class-conditional image generation # ImageNet-64 class label 145 corresponds to king penguins image = pipe(num_inference_steps=1, class_labels=145).images image.save("cd_imagenet64_lpips_onestep_sample_penguin.png") # Multistep sampling, class-conditional image generation # Timesteps can be explicitly specified; the particular timesteps below are from the original Github repo: # https://github.com/openai/consistency_models/blob/main/scripts/launch.sh#L74 image = pipe(num_inference_steps=None, timesteps=[22, 0], class_labels=145).images image.save("cd_imagenet64_lpips_multistep_sample_penguin.png")
- Model type: Consistency model unconditional image generation model, distilled from a diffusion model
- Dataset: ImageNet 64x64
- License: MIT
- Model Description: This model performs unconditional image generation. Its main component is a U-Net, which parameterizes the consistency model. This model was distilled by the Consistency Model authors from an EDM diffusion model, also originally trained by the authors.
- Resources for more information:: Paper, GitHub Repository, Original Model Card
Note: This section is taken from the "Datasets" section of the original model card.
The models that we are making available have been trained on the ILSVRC 2012 subset of ImageNet or on individual categories from LSUN. Here we outline the characteristics of these datasets that influence the behavior of the models:
ILSVRC 2012 subset of ImageNet: This dataset was curated in 2012 and has around a million pictures, each of which belongs to one of 1,000 categories. A significant number of the categories in this dataset are animals, plants, and other naturally occurring objects. Although many photographs include humans, these humans are typically not represented by the class label (for example, the category "Tench, tinca tinca" includes many photographs of individuals holding fish).
LSUN: This dataset was collected in 2015 by a combination of human labeling via Amazon Mechanical Turk and automated data labeling. Both classes that we consider have more than a million images. The dataset creators discovered that when assessed by trained experts, the label accuracy was approximately 90% throughout the entire LSUN dataset. The pictures are gathered from the internet, and those in the cat class often follow a "meme" format. Occasionally, people, including faces, appear in these photographs.
Note: This section is taken from the "Performance" section of the original model card.
These models are intended to generate samples consistent with their training distributions. This has been measured in terms of FID, Inception Score, Precision, and Recall. These metrics all rely on the representations of a pre-trained Inception-V3 model, which was trained on ImageNet, and so is likely to focus more on the ImageNet classes (such as animals) than on other visual features (such as human faces).
Note: This section is taken from the "Intended Use" section of the original model card.
These models are intended to be used for research purposes only. In particular, they can be used as a baseline for generative modeling research, or as a starting point for advancing such research. These models are not intended to be commercially deployed. Additionally, they are not intended to be used to create propaganda or offensive imagery.
Note: This section is taken from the "Limitations" section of the original model card.
These models sometimes produce highly unrealistic outputs, particularly when generating images containing human faces. This may stem from ImageNet's emphasis on non-human objects.
In consistency distillation and training, minimizing LPIPS results in better sample quality, as evidenced by improved FID and Inception scores. However, it also carries the risk of overestimating model performance, because LPIPS uses a VGG network pre-trained on ImageNet, while FID and Inception scores also rely on convolutional neural networks (the Inception network in particular) pre-trained on the same ImageNet dataset. Although these two convolutional neural networks do not share the same architecture and we extract latents from them in substantially different ways, knowledge leakage is still plausible which can undermine the fidelity of FID and Inception scores.
Because ImageNet and LSUN contain images from the internet, they include photos of real people, and the model may have memorized some of the information contained in these photos. However, these images are already publicly available, and existing generative models trained on ImageNet have not demonstrated significant leakage of this information.
- Downloads last month