apolinario commited on
Commit
744a1fc
1 Parent(s): 0098e32

Move all to gallery

Browse files
Files changed (1) hide show
  1. app.py +24 -10
app.py CHANGED
@@ -29,20 +29,34 @@ def text2image_latent(text,steps,width,height,images,diversity):
29
  image_path = f'{temp_dir}/{url}.png'
30
  img.save(f'{temp_dir}/{url}.png')
31
  image_paths.append(image_path)
32
- return(results[0],image_paths)
33
 
34
  def text2image_rudalle(text,aspect,model):
35
  image = rudalle(text,aspect,model)[0]
36
- return(image)
37
 
38
  def text2image_vqgan(text,width,height,style,steps,flavor):
39
  results = vqgan(text,width,height,style,steps,flavor)
40
- return(results)
41
 
42
  def text2image_diffusion(steps_diff, images_diff, weight, clip):
43
  results = diffusion(steps_diff, images_diff, weight, clip)
44
- print(results)
45
- return(results)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46
 
47
  css_mt = {"margin-top": "1em"}
48
 
@@ -85,13 +99,13 @@ with gr.Blocks() as mindseye:
85
  get_image_diffusion = gr.button("Generate Image",css=css_mt)
86
  with gr.Row():
87
  with gr.Tabs():
88
- with gr.TabItem("Image output"):
89
- image = gr.outputs.Image()
90
  with gr.TabItem("Gallery output"):
91
  gallery = gr.Gallery(label="Individual images")
92
 
93
- get_image_latent.click(text2image_latent, inputs=[text,steps,width,height,images,diversity], outputs=[image,gallery])
94
- get_image_rudalle.click(text2image_rudalle, inputs=[text,aspect,model], outputs=image)
95
- get_image_vqgan.click(text2image_vqgan, inputs=[text,width_vq,height_vq,style,steps,flavor],outputs=image)
96
  get_image_diffusion.click(text2image_diffusion, inputs=[steps_diff, images_diff, weight, clip],outputs=gallery)
97
  mindseye.launch()
 
29
  image_path = f'{temp_dir}/{url}.png'
30
  img.save(f'{temp_dir}/{url}.png')
31
  image_paths.append(image_path)
32
+ return(image_paths)
33
 
34
  def text2image_rudalle(text,aspect,model):
35
  image = rudalle(text,aspect,model)[0]
36
+ return([image])
37
 
38
  def text2image_vqgan(text,width,height,style,steps,flavor):
39
  results = vqgan(text,width,height,style,steps,flavor)
40
+ return([results])
41
 
42
  def text2image_diffusion(steps_diff, images_diff, weight, clip):
43
  results = diffusion(steps_diff, images_diff, weight, clip)
44
+ image_paths = []
45
+ image_arrays = []
46
+ for image in results:
47
+ image_str = image[0]
48
+ image_str = image_str.replace("data:image/png;base64,","")
49
+ decoded_bytes = base64.decodebytes(bytes(image_str, "utf-8"))
50
+ img = Image.open(io.BytesIO(decoded_bytes))
51
+ #image_arrays.append(numpy.asarray(img))
52
+ url = shortuuid.uuid()
53
+ temp_dir = './tmp'
54
+ if not os.path.exists(temp_dir):
55
+ os.makedirs(temp_dir, exist_ok=True)
56
+ image_path = f'{temp_dir}/{url}.png'
57
+ img.save(f'{temp_dir}/{url}.png')
58
+ image_paths.append(image_path)
59
+ return(image_paths)
60
 
61
  css_mt = {"margin-top": "1em"}
62
 
 
99
  get_image_diffusion = gr.button("Generate Image",css=css_mt)
100
  with gr.Row():
101
  with gr.Tabs():
102
+ #with gr.TabItem("Image output"):
103
+ # image = gr.outputs.Image()
104
  with gr.TabItem("Gallery output"):
105
  gallery = gr.Gallery(label="Individual images")
106
 
107
+ get_image_latent.click(text2image_latent, inputs=[text,steps,width,height,images,diversity], outputs=gallery)
108
+ get_image_rudalle.click(text2image_rudalle, inputs=[text,aspect,model], outputs=gallery)
109
+ get_image_vqgan.click(text2image_vqgan, inputs=[text,width_vq,height_vq,style,steps,flavor],outputs=gallery)
110
  get_image_diffusion.click(text2image_diffusion, inputs=[steps_diff, images_diff, weight, clip],outputs=gallery)
111
  mindseye.launch()