--- title: Stable Diffusion Playground description: Launch an interactive web app for Stable Diffusion icon: "circle-2" version: EN --- This example deploys a simple web app for Stable Diffusion. You will learn how you can set up an interactive workload for inference -- mounting models from Hugging Face and opening up a port for user inputs. For a more in-depth guide, refer to our [blog post](https://blog.vessl.ai/thin-plate-spline-motion-model-for-image-animation). Try out the Quickstart example with a single click on VESSL Hub. See the completed YAML file and final code for this example. ## What you will do - Host a GPU-accelerated web app built with [Streamlit](https://streamlit.io/) - Mount model checkpoints from [Hugging Face](https://huggingface.co/) - Open up a port to an interactive workload for inference ## Writing the YAML Let's fill in the `stable-diffusion.yml` file. We already learned how you can launch an interactive workload in our [previous](/get-started/gpu-notebook) guide. Let's copy & paste the YAML we wrote for `notebook.yml`. ```yaml name: Stable Diffusion Playground description: An interactive web app for Stable Diffusion resources: cluster: vessl-gcp-oregon preset: gpu-l4-small image: quay.io/vessl-ai/torch:2.1.0-cuda12.2-r3 interactive: jupyter: idle_timeout: 120m max_runtime: 24h ``` Let's mount a [GitHub repo](https://github.com/vessl-ai/hub-model/tree/main/SSD-1B) and import a model checkpoint from Hugging Face. We already learned how you can mount a codebase from our [Quickstart](/get-started/quickstart) guide. VESSL AI comes with a native integration with Hugging Face so you can import models and datasets simply by referencing the link to the Hugging Face repository. Under `import`, let's create a working directory `/model/` and import the [model](https://huggingface.co/VESSL/SSD-1B/tree/main). ```yaml name: Stable Diffusion Playground description: An interactive web app for Stable Diffusion resources: cluster: vessl-gcp-oregon preset: gpu-l4-small image: quay.io/vessl-ai/torch:2.1.0-cuda12.2-r3 import: /code/: git: url: https://github.com/vessl-ai/hub-model ref: main /model/: hf://huggingface.co/VESSL/SSD-1B interactive: jupyter: idle_timeout: 120m max_runtime: 24h ``` The `ports` key expose the workload ports where VESSL AI listens for HTTP requests. This means you will be able to interact with the remote workload -- sending input query and receiving an generated image through port `80` in this case. ```yaml name: Stable Diffusion Playground description: An interactive web app for Stable Diffusion resources: cluster: vessl-gcp-oregon preset: gpu-l4-small image: quay.io/vessl-ai/torch:2.1.0-cuda12.2-r3 import: /code/: git: url: https://github.com/vessl-ai/hub-model ref: main /model/: hf://huggingface.co/VESSL/SSD-1B interactive: jupyter: idle_timeout: 120m max_runtime: 24h ports: - name: streamlit type: http port: 80 ``` Let's install additional Python dependencies with [`requirements.txt`](https://github.com/vessl-ai/hub-model/blob/main/SSD-1B/requirements.txt) and finally run our app [`ssd_1b_streamlit.py`](https://github.com/vessl-ai/hub-model/blob/main/SSD-1B/ssd_1b_streamlit.py). Here, we see how our Streamlit app is using the port we created previously with the `--server.port=80` flag. Through the port, the app receives a user input and generates an image with the Hugging Face model we mounted on `/model/`. ```yaml name: Stable Diffusion Playground description: An interactive web app for Stable Diffusion resources: cluster: vessl-gcp-oregon preset: gpu-l4-small image: quay.io/vessl-ai/torch:2.1.0-cuda12.2-r3 import: /code/: git: url: https://github.com/vessl-ai/hub-model ref: main /model/: hf://huggingface.co/VESSL/SSD-1B run: - command: |- pip install -r requirements.txt streamlit run ssd_1b_streamlit.py --server.port=80 workdir: /code/SSD-1B interactive: max_runtime: 24h jupyter: idle_timeout: 120m ports: - name: streamlit type: http port: 80 ``` ## Running the app Once again, running the workload will guide you to the workload Summary page. ``` vessl run create -f stable-diffusion.yml ``` Under ENDPOINTS, click the `streamlit` link to launch the app. ## Using our web interface You can repeat the same process on the web. Head over to your [Organization](https://vessl.ai), select a project, and create a New run. ## What's next? See how VESSL AI takes care of the infrastructural challenges of fine-tuning a large language model with a custom dataset. Launch an interactive web application for Stable Diffusion