Update image_generator.py
Browse files- image_generator.py +4 -3
image_generator.py
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
from diffusers import StableDiffusionXLPipeline, UNet2DConditionModel, LCMScheduler, DPMSolverMultistepScheduler
|
2 |
import torch
|
3 |
|
4 |
loaded_pipe = None
|
@@ -12,9 +12,10 @@ def load_model(pipe_id, unet_model_id):
|
|
12 |
torch_dtype=torch.float16,
|
13 |
variant="fp16",
|
14 |
)
|
15 |
-
|
|
|
16 |
loaded_pipe = StableDiffusionXLPipeline.from_pretrained(
|
17 |
-
pipe_id, unet=unet, torch_dtype=torch.float16, variant="fp16",
|
18 |
).to("cuda")
|
19 |
loaded_pipe_id = pipe_id
|
20 |
return loaded_pipe
|
|
|
1 |
+
from diffusers import StableDiffusionXLPipeline, AutoencoderKL, UNet2DConditionModel, LCMScheduler, DPMSolverMultistepScheduler
|
2 |
import torch
|
3 |
|
4 |
loaded_pipe = None
|
|
|
12 |
torch_dtype=torch.float16,
|
13 |
variant="fp16",
|
14 |
)
|
15 |
+
vae = AutoencoderKL.from_pretrained("madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16)
|
16 |
+
|
17 |
loaded_pipe = StableDiffusionXLPipeline.from_pretrained(
|
18 |
+
pipe_id, unet=unet, vae=vae, torch_dtype=torch.float16, variant="fp16",
|
19 |
).to("cuda")
|
20 |
loaded_pipe_id = pipe_id
|
21 |
return loaded_pipe
|