--- tags: - stable-diffusion - stable-diffusion-diffusers - text-to-image - diffusers - instruct-pix2pix datasets: - UCSC-VLAA/HQ-Edit --- ## Quick Start Make sure to install the libraries first: ```bash pip install accelerate transformers pip install git+https://github.com/huggingface/diffusers ``` ```python import torch from diffusers import StableDiffusionXLInstructPix2PixPipeline from diffusers.utils import load_image resolution = 768 image = load_image( "https://hf.co/datasets/diffusers/diffusers-images-docs/resolve/main/mountain.png" ).resize((resolution, resolution)) edit_instruction = "Turn sky into a cloudy one" pipe = StableDiffusionXLInstructPix2PixPipeline.from_pretrained( "UCSC-VLAA/HQ-Edit", torch_dtype=torch.float16 ).to("cuda") edited_image = pipe( prompt=edit_instruction, image=image, height=resolution, width=resolution, guidance_scale=3.0, image_guidance_scale=1.5, num_inference_steps=30, ).images[0] edited_image.save("edited_image.png") ```