patrickvonplaten pcuenq HF staff commited on
Commit
d8abfee
1 Parent(s): 1f82a5e

Update LoRA loading (#1)

Browse files

- Update LoRA loading (96233b84bfcfb1553342595483d243d2923b1c44)


Co-authored-by: Pedro Cuenca <pcuenq@users.noreply.huggingface.co>

Files changed (1) hide show
  1. app.py +2 -4
app.py CHANGED
@@ -58,9 +58,7 @@ if TORCH_COMPILE:
58
 
59
  # Load LCM LoRA
60
  pipe.load_lora_weights(
61
- "lcm-sd/lcm-sdxl-lora",
62
- weight_name="lcm_sdxl_lora.safetensors",
63
- #adapter_name="lcm",
64
  use_auth_token=HF_TOKEN,
65
  )
66
 
@@ -144,7 +142,7 @@ with gr.Blocks(css=css) as demo:
144
  from diffusers import DiffusionPipeline, LCMScheduler
145
  pipe = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-base-1.0").to("cuda")
146
  pipe.scheduler = LCMScheduler.from_config(pipe.scheduler.config)
147
- pipe.load_lora_weights("lcm-sd/lcm-sdxl-lora") #yes, it's a normal LoRA
148
 
149
  results = pipe(
150
  prompt="The spirit of a tamagotchi wandering in the city of Vienna",
 
58
 
59
  # Load LCM LoRA
60
  pipe.load_lora_weights(
61
+ "latent-consistency/lcm-lora-sdxl",
 
 
62
  use_auth_token=HF_TOKEN,
63
  )
64
 
 
142
  from diffusers import DiffusionPipeline, LCMScheduler
143
  pipe = DiffusionPipeline.from_pretrained("stabilityai/stable-diffusion-xl-base-1.0").to("cuda")
144
  pipe.scheduler = LCMScheduler.from_config(pipe.scheduler.config)
145
+ pipe.load_lora_weights("latent-consistency/lcm-lora-sdxl") #yes, it's a normal LoRA
146
 
147
  results = pipe(
148
  prompt="The spirit of a tamagotchi wandering in the city of Vienna",