Low VRAM mode docker image of the SVD_XT demo

#23
by macadeliccc - opened

I have made a docker SDK of the SVD_XT demo from the generative-models repository.

You must have docker installed with nvidia GPU capabilities.

The base image is nvidia/cuda:12.2.0-devel-ubuntu20.04 and it also includes the svd_xt.safetensors and the svd_xt_image_decoder.safetensors

To run the image you just need to run

docker pull miniagi/stable_video_diffusion_xt 
docker run --gpus all -p 8501:8501 miniagi/stable_video_diffusion_xt

This will run the video_ sampling.py demo from the generative-models repo.

I have a 4090 and I have tested the Decode t frames at a time (set small if you are low on VRAM) parameter as high as 5 with VRAM usage as high as 22GB

For 16GB cards you can run this with the Decode t frames at a time (set small if you are low on VRAM) parameter set to either 1 or 2.

Double loading models will cause unpredictable behavior including crashes if you exceed VRAM. This issue is present in the base application from stabilityai.

Hey @macadeliccc , Can we run this on 12GB VRAM ? (RTX 3060ti) If yes, can you please tell me what would be ideal configuration settings

@LuciferYagami I havent tested how low the VRAM usage is with only 1 for the decode t frames, but I would say its very unlikely. I will try to put standard svd in the container but its still pushing it.

Thanks for the response. Hoped I could run it with a 3060ti

currently running on rtx 3070 8GB VRAM generating 512x512 pretty good. watch this tutorial : https://www.youtube.com/watch?v=HOVYu2UbgEE

Sign up or log in to comment