Update image_generator.py
Browse files- image_generator.py +2 -4
image_generator.py
CHANGED
@@ -1,9 +1,8 @@
|
|
1 |
-
from diffusers import StableDiffusionXLPipeline, UNet2DConditionModel, LCMScheduler, DPMSolverMultistepScheduler
|
2 |
import torch
|
3 |
|
4 |
loaded_pipe = None
|
5 |
loaded_pipe_id = None
|
6 |
-
vae_model_path = "https://huggingface.co/madebyollin/sdxl-vae-fp16-fix/resolve/main/sdxl_vae.safetensors"
|
7 |
|
8 |
def load_model(pipe_id, unet_model_id):
|
9 |
global loaded_pipe, loaded_pipe_id
|
@@ -13,10 +12,9 @@ def load_model(pipe_id, unet_model_id):
|
|
13 |
torch_dtype=torch.float16,
|
14 |
variant="fp16",
|
15 |
)
|
16 |
-
vae = AutoencoderKL.from_single_file(vae_model_path)
|
17 |
|
18 |
loaded_pipe = StableDiffusionXLPipeline.from_pretrained(
|
19 |
-
pipe_id, unet=unet,
|
20 |
).to("cuda")
|
21 |
loaded_pipe_id = pipe_id
|
22 |
return loaded_pipe
|
|
|
1 |
+
from diffusers import StableDiffusionXLPipeline, UNet2DConditionModel, LCMScheduler, DPMSolverMultistepScheduler
|
2 |
import torch
|
3 |
|
4 |
loaded_pipe = None
|
5 |
loaded_pipe_id = None
|
|
|
6 |
|
7 |
def load_model(pipe_id, unet_model_id):
|
8 |
global loaded_pipe, loaded_pipe_id
|
|
|
12 |
torch_dtype=torch.float16,
|
13 |
variant="fp16",
|
14 |
)
|
|
|
15 |
|
16 |
loaded_pipe = StableDiffusionXLPipeline.from_pretrained(
|
17 |
+
pipe_id, unet=unet, torch_dtype=torch.float16, variant="fp16",
|
18 |
).to("cuda")
|
19 |
loaded_pipe_id = pipe_id
|
20 |
return loaded_pipe
|