Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -136,7 +136,7 @@ def sign_message(message, key):
|
|
136 |
signed_hash = base64.b64encode(hmac_digest).decode()
|
137 |
return signed_hash
|
138 |
|
139 |
-
def run(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_prompt=NEGATIVE_PROMPT, guidance_scale=4.0, randomize_seed=True, seed=42, tpu_inference=False, do_img2img=False, init_image=None, image2image_resize=False, image2image_strength=0, progress=gr.Progress(track_tqdm=True)):
|
140 |
if init_image is None:
|
141 |
do_img2img = False
|
142 |
|
@@ -150,11 +150,11 @@ def run(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_promp
|
|
150 |
prompt = prompt.replace("!", " ").replace("\n", " ") # remote endpoint unsupported
|
151 |
if do_img2img:
|
152 |
init_image = codecs.encode(pickle.dumps(init_image, protocol=pickle.HIGHEST_PROTOCOL), "base64").decode('latin1')
|
153 |
-
return tpu_inference_api(prompt, radio, preset, h, w, negative_prompt, guidance_scale, randomize_seed, seed, do_img2img, init_image, image2image_strength)
|
154 |
else:
|
155 |
-
return tpu_inference_api(prompt, radio, preset, h, w, negative_prompt, guidance_scale, randomize_seed, seed)
|
156 |
|
157 |
-
return zero_inference_api(prompt, radio, preset, h, w, negative_prompt, guidance_scale, randomize_seed, seed, do_img2img, init_image, image2image_strength)
|
158 |
|
159 |
@spaces.GPU
|
160 |
def zero_inference_api(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_prompt=NEGATIVE_PROMPT, guidance_scale=4.0, randomize_seed=True, seed=42, do_img2img=False, init_image=None, image2image_strength=0, inference_steps=25, progress=gr.Progress(track_tqdm=True)):
|
@@ -267,7 +267,7 @@ with gr.Blocks(theme=theme) as demo:
|
|
267 |
image2image_resize = gr.Checkbox(label="Resize input image", value=False, visible=False)
|
268 |
image2image_strength = gr.Slider(minimum=0.0, maximum=1.0, step=0.01, label="Noising strength", value=0.7, visible=False)
|
269 |
|
270 |
-
with gr.Column(
|
271 |
output = gr.Image(type="filepath", interactive=False)
|
272 |
|
273 |
gr.Examples(fn=run, examples=["mayano_top_gun_\(umamusume\), 1girl, rurudo", "sho (sho lwlw),[[[ohisashiburi]]],fukuro daizi,tianliang duohe fangdongye,[daidai ookami],year_2023, (wariza), depth of field, official_art"], inputs=prompt, outputs=[output, seed], cache_examples="lazy")
|
|
|
136 |
signed_hash = base64.b64encode(hmac_digest).decode()
|
137 |
return signed_hash
|
138 |
|
139 |
+
def run(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_prompt=NEGATIVE_PROMPT, guidance_scale=4.0, randomize_seed=True, seed=42, tpu_inference=False, do_img2img=False, init_image=None, image2image_resize=False, image2image_strength=0, inference_steps=25, progress=gr.Progress(track_tqdm=True)):
|
140 |
if init_image is None:
|
141 |
do_img2img = False
|
142 |
|
|
|
150 |
prompt = prompt.replace("!", " ").replace("\n", " ") # remote endpoint unsupported
|
151 |
if do_img2img:
|
152 |
init_image = codecs.encode(pickle.dumps(init_image, protocol=pickle.HIGHEST_PROTOCOL), "base64").decode('latin1')
|
153 |
+
return tpu_inference_api(prompt, radio, preset, h, w, negative_prompt, guidance_scale, randomize_seed, seed, do_img2img, init_image, image2image_strength, inference_steps=inference_steps)
|
154 |
else:
|
155 |
+
return tpu_inference_api(prompt, radio, preset, h, w, negative_prompt, guidance_scale, randomize_seed, seed, inference_steps=inference_steps)
|
156 |
|
157 |
+
return zero_inference_api(prompt, radio, preset, h, w, negative_prompt, guidance_scale, randomize_seed, seed, do_img2img, init_image, image2image_strength, inference_steps=inference_steps)
|
158 |
|
159 |
@spaces.GPU
|
160 |
def zero_inference_api(prompt, radio="model-v2", preset=PRESET_Q, h=1216, w=832, negative_prompt=NEGATIVE_PROMPT, guidance_scale=4.0, randomize_seed=True, seed=42, do_img2img=False, init_image=None, image2image_strength=0, inference_steps=25, progress=gr.Progress(track_tqdm=True)):
|
|
|
267 |
image2image_resize = gr.Checkbox(label="Resize input image", value=False, visible=False)
|
268 |
image2image_strength = gr.Slider(minimum=0.0, maximum=1.0, step=0.01, label="Noising strength", value=0.7, visible=False)
|
269 |
|
270 |
+
with gr.Column():
|
271 |
output = gr.Image(type="filepath", interactive=False)
|
272 |
|
273 |
gr.Examples(fn=run, examples=["mayano_top_gun_\(umamusume\), 1girl, rurudo", "sho (sho lwlw),[[[ohisashiburi]]],fukuro daizi,tianliang duohe fangdongye,[daidai ookami],year_2023, (wariza), depth of field, official_art"], inputs=prompt, outputs=[output, seed], cache_examples="lazy")
|