JoPmt commited on
Commit
6a9055b
1 Parent(s): 2f1b9d9

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +15 -14
app.py CHANGED
@@ -1,23 +1,24 @@
1
  import torch
2
-
3
  from diffusers import UniDiffuserPipeline
4
  from diffusers.utils import load_image
 
 
5
 
6
- device = "cuda"
7
- model_id_or_path = "thu-ml/unidiffuser-v1"
8
- pipe = UniDiffuserPipeline.from_pretrained(model_id_or_path, torch_dtype=torch.float16)
9
- pipe.to(device)
10
 
11
- # Image variation can be performed with an image-to-text generation followed by a text-to-image generation:
12
- # 1. Image-to-text generation
13
- image_url = "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unidiffuser/unidiffuser_example_image.jpg"
14
  init_image = load_image(image_url).resize((512, 512))
15
 
16
- sample = pipe(image=init_image, num_inference_steps=20, guidance_scale=8.0)
17
  i2t_text = sample.text[0]
18
- print(i2t_text)
19
 
20
- # 2. Text-to-image generation
21
- sample = pipe(prompt=i2t_text, num_inference_steps=20, guidance_scale=8.0)
22
- final_image = sample.images[0]
23
- final_image.save("unidiffuser_image_variation_sample.png")
 
 
 
 
 
1
  import torch
2
+ import gradio as gr
3
  from diffusers import UniDiffuserPipeline
4
  from diffusers.utils import load_image
5
+ from accelerate import Accelerator
6
+ Accelerator = Accelerator(cpu=True)
7
 
8
+ pipe = accelerator.prepare(UniDiffuserPipeline.from_pretrained("thu-ml/unidiffuser-v1", torch_dtype=torch.bfloat16))
9
+ pipe = accelerator.prepare(pipe.to("cpu"))
 
 
10
 
11
+ def plex(image_url,stips)
 
 
12
  init_image = load_image(image_url).resize((512, 512))
13
 
14
+ sample = pipe(image=init_image, num_inference_steps=stips, guidance_scale=8.0)
15
  i2t_text = sample.text[0]
 
16
 
17
+ sample = pipe(prompt=i2t_text, num_inference_steps=stips, guidance_scale=8.0)
18
+ for i, imge in enumerate(sample["images"]):
19
+ apol.append(imge)
20
+ return apol
21
+
22
+ iface = gr.Interface(fn=plex, inputs=[gr.Image(label="img",type="filepath"), gr.Slider(label="num inference steps", minimum=1, step=1, maximum=5, value=5)], outputs=gr.Gallery(label="out", columns=2))
23
+ iface.queue(max_size=1)
24
+ iface.launch(max_threads=1)