Diffusers documentation

Loading

You are viewing v0.11.0 version. A newer version v0.32.1 is available.
Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

Loading

A core premise of the diffusers library is to make diffusion models as accessible as possible. Accessibility is therefore achieved by providing an API to load complete diffusion pipelines as well as individual components with a single line of code.

In the following we explain in-detail how to easily load:

Loading pipelines

The DiffusionPipeline class is the easiest way to access any diffusion model that is available on the Hub. Let’s look at an example on how to download CompVis’ Latent Diffusion model.

from diffusers import DiffusionPipeline

repo_id = "CompVis/ldm-text2im-large-256"
ldm = DiffusionPipeline.from_pretrained(repo_id)

Here DiffusionPipeline automatically detects the correct pipeline (i.e. LDMTextToImagePipeline), downloads and caches all required configuration and weight files (if not already done so), and finally returns a pipeline instance, called ldm. The pipeline instance can then be called using LDMTextToImagePipeline.call() (i.e., ldm("image of a astronaut riding a horse")) for text-to-image generation.

Instead of using the generic DiffusionPipeline class for loading, you can also load the appropriate pipeline class directly. The code snippet above yields the same instance as when doing:

from diffusers import LDMTextToImagePipeline

repo_id = "CompVis/ldm-text2im-large-256"
ldm = LDMTextToImagePipeline.from_pretrained(repo_id)

Diffusion pipelines like LDMTextToImagePipeline often consist of multiple components. These components can be both parameterized models, such as "unet", "vqvae" and “bert”, tokenizers or schedulers. These components can interact in complex ways with each other when using the pipeline in inference, e.g. for LDMTextToImagePipeline or StableDiffusionPipeline the inference call is explained here. The purpose of the pipeline classes is to wrap the complexity of these diffusion systems and give the user an easy-to-use API while staying flexible for customization, as will be shown later.

Loading pipelines that require access request

Due to the capabilities of diffusion models to generate extremely realistic images, there is a certain danger that such models might be misused for unwanted applications, e.g. generating pornography or violent images. In order to minimize the possibility of such unsolicited use cases, some of the most powerful diffusion models require users to acknowledge a license before being able to use the model. If the user does not agree to the license, the pipeline cannot be downloaded. If you try to load runwayml/stable-diffusion-v1-5 the same way as done previously:

from diffusers import DiffusionPipeline

repo_id = "runwayml/stable-diffusion-v1-5"
stable_diffusion = DiffusionPipeline.from_pretrained(repo_id)

it will only work if you have both click-accepted the license on the model card and are logged into the Hugging Face Hub. Otherwise you will get an error message such as the following:

OSError: runwayml/stable-diffusion-v1-5 is not a local folder and is not a valid model identifier listed on 'https://huggingface.co/models'
If this is a private repository, make sure to pass a token having permission to this repo with `use_auth_token` or log in with `huggingface-cli login`

Therefore, we need to make sure to click-accept the license. You can do this by simply visiting the model card and clicking on “Agree and access repository”:



Second, you need to login with your access token:

huggingface-cli login

before trying to load the model. Or alternatively, you can pass your access token directly via the flag use_auth_token. In this case you do not need to run huggingface-cli login before:

from diffusers import DiffusionPipeline

repo_id = "runwayml/stable-diffusion-v1-5"
stable_diffusion = DiffusionPipeline.from_pretrained(repo_id, use_auth_token="<your-access-token>")

The final option to use pipelines that require access without having to rely on the Hugging Face Hub is to load the pipeline locally as explained in the next section.

Loading pipelines locally

If you prefer to have complete control over the pipeline and its corresponding files or, as said before, if you want to use pipelines that require an access request without having to be connected to the Hugging Face Hub, we recommend loading pipelines locally.

To load a diffusion pipeline locally, you first need to manually download the whole folder structure on your local disk and then pass a local path to the DiffusionPipeline.from_pretrained(). Let’s again look at an example for CompVis’ Latent Diffusion model.

First, you should make use of git-lfs to download the whole folder structure that has been uploaded to the model repository:

git lfs install
git clone https://huggingface.co/runwayml/stable-diffusion-v1-5

The command above will create a local folder called ./stable-diffusion-v1-5 on your disk. Now, all you have to do is to simply pass the local folder path to from_pretrained:

from diffusers import DiffusionPipeline

repo_id = "./stable-diffusion-v1-5"
stable_diffusion = DiffusionPipeline.from_pretrained(repo_id)

If repo_id is a local path, as it is the case here, DiffusionPipeline.from_pretrained() will automatically detect it and therefore not try to download any files from the Hub. While we usually recommend to load weights directly from the Hub to be certain to stay up to date with the newest changes, loading pipelines locally should be preferred if one wants to stay anonymous, self-contained applications, etc…

Loading customized pipelines

Advanced users that want to load customized versions of diffusion pipelines can do so by swapping any of the default components, e.g. the scheduler, with other scheduler classes. A classical use case of this functionality is to swap the scheduler. Stable Diffusion v1-5 uses the PNDMScheduler by default which is generally not the most performant scheduler. Since the release of stable diffusion, multiple improved schedulers have been published. To use those, the user has to manually load their preferred scheduler and pass it into DiffusionPipeline.from_pretrained().

E.g. to use EulerDiscreteScheduler or DPMSolverMultistepScheduler to have a better quality vs. generation speed trade-off for inference, one could load them as follows:

from diffusers import DiffusionPipeline, EulerDiscreteScheduler, DPMSolverMultistepScheduler

repo_id = "runwayml/stable-diffusion-v1-5"

scheduler = EulerDiscreteScheduler.from_pretrained(repo_id, subfolder="scheduler")
# or
# scheduler = DPMSolverMultistepScheduler.from_pretrained(repo_id, subfolder="scheduler")

stable_diffusion = DiffusionPipeline.from_pretrained(repo_id, scheduler=scheduler)

Three things are worth paying attention to here.

Not only the scheduler components can be customized for diffusion pipelines; in theory, all components of a pipeline can be customized. In practice, however, it often only makes sense to switch out a component that has compatible alternatives to what the pipeline expects. Many scheduler classes are compatible with each other as can be seen here. This is not always the case for other components, such as the "unet".

One special case that can also be customized is the "safety_checker" of stable diffusion. If you believe the safety checker doesn’t serve you any good, you can simply disable it by passing None:

from diffusers import DiffusionPipeline, EulerDiscreteScheduler, DPMSolverMultistepScheduler

stable_diffusion = DiffusionPipeline.from_pretrained(repo_id, safety_checker=None)

Another common use case is to reuse the same components in multiple pipelines, e.g. the weights and configurations of "runwayml/stable-diffusion-v1-5" can be used for both StableDiffusionPipeline and StableDiffusionImg2ImgPipeline and we might not want to use the exact same weights into RAM twice. In this case, customizing all the input instances would help us to only load the weights into RAM once:

from diffusers import StableDiffusionPipeline, StableDiffusionImg2ImgPipeline

model_id = "runwayml/stable-diffusion-v1-5"
stable_diffusion_txt2img = StableDiffusionPipeline.from_pretrained(model_id)

components = stable_diffusion_txt2img.components

# weights are not reloaded into RAM
stable_diffusion_img2img = StableDiffusionImg2ImgPipeline(**components)

Note how the above code snippet makes use of DiffusionPipeline.components.

How does loading work?

As a class method, DiffusionPipeline.from_pretrained() is responsible for two things:

  • Download the latest version of the folder structure required to run the repo_id with diffusers and cache them. If the latest folder structure is available in the local cache, DiffusionPipeline.from_pretrained() will simply reuse the cache and not re-download the files.
  • Load the cached weights into the correct pipeline class – one of the officially supported pipeline classes - and return an instance of the class. The correct pipeline class is thereby retrieved from the model_index.json file.

The underlying folder structure of diffusion pipelines correspond 1-to-1 to their corresponding class instances, e.g. LDMTextToImagePipeline for CompVis/ldm-text2im-large-256 This can be understood better by looking at an example. Let’s print out pipeline class instance pipeline we just defined:

from diffusers import DiffusionPipeline

repo_id = "CompVis/ldm-text2im-large-256"
ldm = DiffusionPipeline.from_pretrained(repo_id)
print(ldm)

Output:

LDMTextToImagePipeline {
  "bert": [
    "latent_diffusion",
    "LDMBertModel"
  ],
  "scheduler": [
    "diffusers",
    "DDIMScheduler"
  ],
  "tokenizer": [
    "transformers",
    "BertTokenizer"
  ],
  "unet": [
    "diffusers",
    "UNet2DConditionModel"
  ],
  "vqvae": [
    "diffusers",
    "AutoencoderKL"
  ]
}

First, we see that the official pipeline is the LDMTextToImagePipeline, and second we see that the LDMTextToImagePipeline consists of 5 components:

Let’s now compare the pipeline instance to the folder structure of the model repository CompVis/ldm-text2im-large-256. Looking at the folder structure of CompVis/ldm-text2im-large-256 on the Hub, we can see it matches 1-to-1 the printed out instance of LDMTextToImagePipeline above:

.
├── bert
│   ├── config.json
│   └── pytorch_model.bin
├── model_index.json
├── scheduler
│   └── scheduler_config.json
├── tokenizer
│   ├── special_tokens_map.json
│   ├── tokenizer_config.json
│   └── vocab.txt
├── unet
│   ├── config.json
│   └── diffusion_pytorch_model.bin
└── vqvae
    ├── config.json
    └── diffusion_pytorch_model.bin

As we can see each attribute of the instance of LDMTextToImagePipeline has its configuration and possibly weights defined in a subfolder that is called exactly like the class attribute ("bert", "scheduler", "tokenizer", "unet", "vqvae"). Importantly, every pipeline expects a model_index.json file that tells the DiffusionPipeline both:

  • which pipeline class should be loaded, and
  • what sub-classes from which library are stored in which subfolders

In the case of CompVis/ldm-text2im-large-256 the model_index.json is therefore defined as follows:

{
  "_class_name": "LDMTextToImagePipeline",
  "_diffusers_version": "0.0.4",
  "bert": [
    "latent_diffusion",
    "LDMBertModel"
  ],
  "scheduler": [
    "diffusers",
    "DDIMScheduler"
  ],
  "tokenizer": [
    "transformers",
    "BertTokenizer"
  ],
  "unet": [
    "diffusers",
    "UNet2DConditionModel"
  ],
  "vqvae": [
    "diffusers",
    "AutoencoderKL"
  ]
}
  • _class_name tells DiffusionPipeline which pipeline class should be loaded.
  • _diffusers_version can be useful to know under which diffusers version this model was created.
  • Every component of the pipeline is then defined under the form:
"name" : [
  "library",
  "class"
]
  • The "name" field corresponds both to the name of the subfolder in which the configuration and weights are stored as well as the attribute name of the pipeline class (as can be seen here and here
  • The "library" field corresponds to the name of the library, e.g. diffusers or transformers from which the "class" should be loaded
  • The "class" field corresponds to the name of the class, e.g. BertTokenizer or UNet2DConditionModel

Loading models

Models as defined under src/diffusers/models can be loaded via the ModelMixin.from_pretrained() function. The API is very similar the DiffusionPipeline.from_pretrained() and works in the same way:

  • Download the latest version of the model weights and configuration with diffusers and cache them. If the latest files are available in the local cache, ModelMixin.from_pretrained() will simply reuse the cache and not re-download the files.
  • Load the cached weights into the defined model class - one of the existing model classes - and return an instance of the class.

In constrast to DiffusionPipeline.from_pretrained(), models rely on fewer files that usually don’t require a folder structure, but just a diffusion_pytorch_model.bin and config.json file.

Let’s look at an example:

from diffusers import UNet2DConditionModel

repo_id = "CompVis/ldm-text2im-large-256"
model = UNet2DConditionModel.from_pretrained(repo_id, subfolder="unet")

Note how we have to define the subfolder="unet" argument to tell ModelMixin.from_pretrained() that the model weights are located in a subfolder of the repository.

As explained in Loading customized pipelines, one can pass a loaded model to a diffusion pipeline, via DiffusionPipeline.from_pretrained():

from diffusers import DiffusionPipeline

repo_id = "CompVis/ldm-text2im-large-256"
ldm = DiffusionPipeline.from_pretrained(repo_id, unet=model)

If the model files can be found directly at the root level, which is usually only the case for some very simple diffusion models, such as google/ddpm-cifar10-32, we don’t need to pass a subfolder argument:

from diffusers import UNet2DModel

repo_id = "google/ddpm-cifar10-32"
model = UNet2DModel.from_pretrained(repo_id)

Loading schedulers

Schedulers rely on SchedulerMixin.from_pretrained(). Schedulers are not parameterized or trained, but instead purely defined by a configuration file. For consistency, we use the same method name as we do for models or pipelines, but no weights are loaded in this case.

In constrast to pipelines or models, loading schedulers does not consume any significant amount of memory and the same configuration file can often be used for a variety of different schedulers. For example, all of:

are compatible with StableDiffusionPipeline and therefore the same scheduler configuration file can be loaded in any of those classes:

from diffusers import StableDiffusionPipeline
from diffusers import (
    DDPMScheduler,
    DDIMScheduler,
    PNDMScheduler,
    LMSDiscreteScheduler,
    EulerDiscreteScheduler,
    EulerAncestralDiscreteScheduler,
    DPMSolverMultistepScheduler,
)

repo_id = "runwayml/stable-diffusion-v1-5"

ddpm = DDPMScheduler.from_pretrained(repo_id, subfolder="scheduler")
ddim = DDIMScheduler.from_pretrained(repo_id, subfolder="scheduler")
pndm = PNDMScheduler.from_pretrained(repo_id, subfolder="scheduler")
lms = LMSDiscreteScheduler.from_pretrained(repo_id, subfolder="scheduler")
euler_anc = EulerAncestralDiscreteScheduler.from_pretrained(repo_id, subfolder="scheduler")
euler = EulerDiscreteScheduler.from_pretrained(repo_id, subfolder="scheduler")
dpm = DPMSolverMultistepScheduler.from_pretrained(repo_id, subfolder="scheduler")

# replace `dpm` with any of `ddpm`, `ddim`, `pndm`, `lms`, `euler`, `euler_anc`
pipeline = StableDiffusionPipeline.from_pretrained(repo_id, scheduler=dpm)