Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -80,7 +80,6 @@ def update_selection(evt: gr.SelectData, width, height):
|
|
80 |
height,
|
81 |
)
|
82 |
|
83 |
-
@spaces.GPU(duration=45)
|
84 |
def generate_image(prompt_mash, steps, seed, cfg_scale, width, height, lora_scale, progress):
|
85 |
if torch.cuda.is_available():
|
86 |
pipe.to("cuda")
|
@@ -196,7 +195,7 @@ def add_custom_lora(custom_lora):
|
|
196 |
def remove_custom_lora():
|
197 |
return gr.update(visible=False), gr.update(visible=False), gr.update(), "", None, ""
|
198 |
|
199 |
-
@spaces.GPU(duration=
|
200 |
def run_lora(prompt, image_input, image_strength, cfg_scale, steps, selected_index, randomize_seed, seed, width, height, lora_scale, progress=gr.Progress(track_tqdm=True)):
|
201 |
if selected_index is None:
|
202 |
raise gr.Error("You must select a LoRA before proceeding.")
|
|
|
80 |
height,
|
81 |
)
|
82 |
|
|
|
83 |
def generate_image(prompt_mash, steps, seed, cfg_scale, width, height, lora_scale, progress):
|
84 |
if torch.cuda.is_available():
|
85 |
pipe.to("cuda")
|
|
|
195 |
def remove_custom_lora():
|
196 |
return gr.update(visible=False), gr.update(visible=False), gr.update(), "", None, ""
|
197 |
|
198 |
+
@spaces.GPU(duration=40)
|
199 |
def run_lora(prompt, image_input, image_strength, cfg_scale, steps, selected_index, randomize_seed, seed, width, height, lora_scale, progress=gr.Progress(track_tqdm=True)):
|
200 |
if selected_index is None:
|
201 |
raise gr.Error("You must select a LoRA before proceeding.")
|