import gradio as gr import numpy as np import ast from matplotlib.patches import Polygon from matplotlib.collections import PatchCollection import matplotlib.pyplot as plt from utils.parse import filter_boxes from generation import run as run_ours from baseline import run as run_baseline import torch print(f"Is CUDA available: {torch.cuda.is_available()}") if torch.cuda.is_available(): print(f"CUDA device: {torch.cuda.get_device_name(torch.cuda.current_device())}") box_scale = (512, 512) size = box_scale bg_prompt_text = "Background prompt: " simplified_prompt = """You are an intelligent bounding box generator. I will provide you with a caption for a photo, image, or painting. Your task is to generate the bounding boxes for the objects mentioned in the caption, along with a background prompt describing the scene. The images are of size 512x512, and the bounding boxes should not overlap or go beyond the image boundaries. Each bounding box should be in the format of (object name, [top-left x coordinate, top-left y coordinate, box width, box height]) and include exactly one object. Do not put objects that are already provided in the bounding boxes into the background prompt. If needed, you can make reasonable guesses. Generate the object descriptions and background prompts in English even if the caption might not be in English. Please refer to the example below for the desired format. Caption: A realistic image of landscape scene depicting a green car parking on the left of a blue truck, with a red air balloon and a bird in the sky Objects: [('a green car', [21, 181, 211, 159]), ('a blue truck', [269, 181, 209, 160]), ('a red air balloon', [66, 8, 145, 135]), ('a bird', [296, 42, 143, 100])] Background prompt: A realistic image of a landscape scene Caption: A watercolor painting of a wooden table in the living room with an apple on it Objects: [('a wooden table', [65, 243, 344, 206]), ('a apple', [206, 306, 81, 69])] Background prompt: A watercolor painting of a living room Caption: A watercolor painting of two pandas eating bamboo in a forest Objects: [('a panda eating bambooo', [30, 171, 212, 226]), ('a panda eating bambooo', [264, 173, 222, 221])] Background prompt: A watercolor painting of a forest Caption: A realistic image of four skiers standing in a line on the snow near a palm tree Objects: [('a skier', [5, 152, 139, 168]), ('a skier', [278, 192, 121, 158]), ('a skier', [148, 173, 124, 155]), ('a palm tree', [404, 180, 103, 180])] Background prompt: A realistic image of an outdoor scene with snow Caption: An oil painting of a pink dolphin jumping on the left of a steam boat on the sea Objects: [('a steam boat', [232, 225, 257, 149]), ('a jumping pink dolphin', [21, 249, 189, 123])] Background prompt: An oil painting of the sea Caption: A realistic image of a cat playing with a dog in a park with flowers Objects: [('a playful cat', [51, 67, 271, 324]), ('a playful dog', [302, 119, 211, 228])] Background prompt: A realistic image of a park with flowers Caption: 一个客厅场景的油画,墙上挂着电视,电视下面是一个柜子,柜子上有一个花瓶。 Objects: [('a tv', [88, 85, 335, 203]), ('a cabinet', [57, 308, 404, 201]), ('a flower vase', [166, 222, 92, 108])] Background prompt: An oil painting of a living room scene Caption: {prompt} Objects: """ prompt_placeholder = "A realistic photo of a gray cat and an orange dog on the grass." layout_placeholder = """Caption: A realistic photo of a gray cat and an orange dog on the grass. Objects: [('a gray cat', [67, 243, 120, 126]), ('an orange dog', [265, 193, 190, 210])] Background prompt: A realistic photo of a grassy area.""" def get_lmd_prompt(prompt): if prompt == "": prompt = prompt_placeholder return simplified_prompt.format(prompt=prompt) def get_layout_image(response): if response == "": response = layout_placeholder gen_boxes, bg_prompt = parse_input(response) fig = plt.figure(figsize=(8, 8)) # https://stackoverflow.com/questions/7821518/save-plot-to-numpy-array show_boxes(gen_boxes, bg_prompt) # If we haven't already shown or saved the plot, then we need to # draw the figure first... fig.canvas.draw() # Now we can save it to a numpy array. data = np.frombuffer(fig.canvas.tostring_rgb(), dtype=np.uint8) data = data.reshape(fig.canvas.get_width_height()[::-1] + (3,)) plt.clf() return data def get_layout_image_gallery(response): return [get_layout_image(response)] def get_ours_image(response, seed, fg_seed_start, fg_blending_ratio=0.1, frozen_step_ratio=0.4, gligen_scheduled_sampling_beta=0.3, show_so_imgs=False, scale_boxes=False, gallery=None): if response == "": response = layout_placeholder gen_boxes, bg_prompt = parse_input(response) gen_boxes = filter_boxes(gen_boxes, scale_boxes=scale_boxes) spec = { # prompt is unused 'prompt': '', 'gen_boxes': gen_boxes, 'bg_prompt': bg_prompt } image_np, so_img_list = run_ours( spec, bg_seed=seed, fg_seed_start=fg_seed_start, fg_blending_ratio=fg_blending_ratio,frozen_step_ratio=frozen_step_ratio, gligen_scheduled_sampling_beta=gligen_scheduled_sampling_beta) images = [image_np] if show_so_imgs: images.extend([np.asarray(so_img) for so_img in so_img_list]) return images def get_baseline_image(prompt, seed): if prompt == "": prompt = prompt_placeholder image_np = run_baseline(prompt, bg_seed=seed) return [image_np] def parse_input(text=None): try: if "Objects: " in text: text = text.split("Objects: ")[1] text_split = text.split(bg_prompt_text) if len(text_split) == 2: gen_boxes, bg_prompt = text_split gen_boxes = ast.literal_eval(gen_boxes) bg_prompt = bg_prompt.strip() except Exception as e: raise gr.Error(f"response format invalid: {e} (text: {text})") return gen_boxes, bg_prompt def draw_boxes(anns): ax = plt.gca() ax.set_autoscale_on(False) polygons = [] color = [] for ann in anns: c = (np.random.random((1, 3))*0.6+0.4) [bbox_x, bbox_y, bbox_w, bbox_h] = ann['bbox'] poly = [[bbox_x, bbox_y], [bbox_x, bbox_y+bbox_h], [bbox_x+bbox_w, bbox_y+bbox_h], [bbox_x+bbox_w, bbox_y]] np_poly = np.array(poly).reshape((4, 2)) polygons.append(Polygon(np_poly)) color.append(c) # print(ann) name = ann['name'] if 'name' in ann else str(ann['category_id']) ax.text(bbox_x, bbox_y, name, style='italic', bbox={'facecolor': 'white', 'alpha': 0.7, 'pad': 5}) p = PatchCollection(polygons, facecolor='none', edgecolors=color, linewidths=2) ax.add_collection(p) def show_boxes(gen_boxes, bg_prompt=None): anns = [{'name': gen_box[0], 'bbox': gen_box[1]} for gen_box in gen_boxes] # White background (to allow line to show on the edge) I = np.ones((size[0]+4, size[1]+4, 3), dtype=np.uint8) * 255 plt.imshow(I) plt.axis('off') if bg_prompt is not None: ax = plt.gca() ax.text(0, 0, bg_prompt, style='italic', bbox={'facecolor': 'white', 'alpha': 0.7, 'pad': 5}) c = np.zeros((1, 3)) [bbox_x, bbox_y, bbox_w, bbox_h] = (0, 0, size[1], size[0]) poly = [[bbox_x, bbox_y], [bbox_x, bbox_y+bbox_h], [bbox_x+bbox_w, bbox_y+bbox_h], [bbox_x+bbox_w, bbox_y]] np_poly = np.array(poly).reshape((4, 2)) polygons = [Polygon(np_poly)] color = [c] p = PatchCollection(polygons, facecolor='none', edgecolors=color, linewidths=2) ax.add_collection(p) draw_boxes(anns) duplicate_html = 'Duplicate Space' with gr.Blocks( title="LLM-grounded Diffusion: Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models" ) as g: gr.HTML(f"""

LLM-grounded Diffusion: Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models

LLM + Stable Diffusion => better prompt understanding in text2image generation 🤩

Project Page | 5-minute Blog Post | ArXiv Paper | Github | Cite our work if our ideas inspire you.

Tips:

1. If ChatGPT doesn't generate layout, add/remove the trailing space (added by default) and/or use GPT-4.

2. You can perform multi-round specification by giving ChatGPT follow-up requests (e.g., make the object boxes bigger).

3. You can also try prompts in Simplified Chinese. If you want to try prompts in another language, translate the first line of last example to your language.

4. Duplicate this space and add GPU to skip the queue and run our model faster. {duplicate_html}


Implementation note: In this demo, we replace the attention manipulation in our layout-guided Stable Diffusion described in our paper with GLIGEN due to much faster inference speed (FlashAttention supported, no backprop needed during inference). Compared to vanilla GLIGEN, we have better coherence. Other parts of text-to-image pipeline, including single object generation and SAM, remain the same. The settings and examples in the prompt are simplified in this demo.

""") with gr.Tab("Stage 1. Image Prompt to ChatGPT"): with gr.Row(): with gr.Column(scale=1): prompt = gr.Textbox(lines=2, label="Prompt for Layout Generation", placeholder=prompt_placeholder) generate_btn = gr.Button("Generate Prompt") with gr.Column(scale=1): output = gr.Textbox(label="Paste this into ChatGPT (GPT-4 preferred; on Mac, click text and press Command+A and Command+C to copy all)") generate_btn.click(fn=get_lmd_prompt, inputs=prompt, outputs=output, api_name="get_lmd_prompt") # with gr.Tab("(Optional) Visualize ChatGPT-generated Layout"): # with gr.Row(): # with gr.Column(scale=1): # response = gr.Textbox(lines=5, label="Paste ChatGPT response here", placeholder=layout_placeholder) # visualize_btn = gr.Button("Visualize Layout") # with gr.Column(scale=1): # output = gr.Image(shape=(512, 512), elem_classes="img", elem_id="img", css="img {width: 300px}") # visualize_btn.click(fn=get_layout_image, inputs=response, outputs=output, api_name="visualize-layout") with gr.Tab("Stage 2 (New). Layout to Image generation"): with gr.Row(): with gr.Column(scale=1): response = gr.Textbox(lines=5, label="Paste ChatGPT response here (no original caption needed)", placeholder=layout_placeholder) visualize_btn = gr.Button("Visualize Layout") generate_btn = gr.Button("Generate Image from Layout", variant='primary') with gr.Accordion("Advanced options", open=False): seed = gr.Slider(0, 10000, value=0, step=1, label="Seed") fg_seed_start = gr.Slider(0, 10000, value=20, step=1, label="Seed for foreground variation") fg_blending_ratio = gr.Slider(0, 1, value=0.1, step=0.01, label="Variations added to foreground for single object generation (0: no variation, 1: max variation)") frozen_step_ratio = gr.Slider(0, 1, value=0.4, step=0.1, label="Foreground frozen steps ratio (higher: preserve object attributes; lower: higher coherence; set to 0: (almost) equivalent to vanilla GLIGEN except details)") gligen_scheduled_sampling_beta = gr.Slider(0, 1, value=0.3, step=0.1, label="GLIGEN guidance steps ratio (the beta value)") show_so_imgs = gr.Checkbox(label="Show annotated single object generations", show_label=False) with gr.Column(scale=1): gallery = gr.Gallery( label="Generated image", show_label=False, elem_id="gallery" ).style(columns=[1], rows=[1], object_fit="contain", preview=True) visualize_btn.click(fn=get_layout_image_gallery, inputs=response, outputs=gallery, api_name="visualize-layout") generate_btn.click(fn=get_ours_image, inputs=[response, seed, fg_seed_start, fg_blending_ratio, frozen_step_ratio, gligen_scheduled_sampling_beta, show_so_imgs], outputs=gallery, api_name="layout-to-image") with gr.Tab("Baseline: Stable Diffusion"): with gr.Row(): with gr.Column(scale=1): sd_prompt = gr.Textbox(lines=2, label="Prompt for baseline SD", placeholder=prompt_placeholder) generate_btn = gr.Button("Generate") with gr.Accordion("Advanced options", open=False): seed = gr.Slider(0, 10000, value=0, step=1, label="Seed") # with gr.Column(scale=1): # output = gr.Image(shape=(512, 512), elem_classes="img", elem_id="img") with gr.Column(scale=1): gallery = gr.Gallery( label="Generated image", show_label=False, elem_id="gallery2" ).style(columns=[1], rows=[1], object_fit="contain", preview=True) generate_btn.click(fn=get_baseline_image, inputs=[sd_prompt, seed], outputs=gallery, api_name="baseline") g.launch()