# Tune-A-Video This repository is the official implementation of [Tune-A-Video](https://arxiv.org/abs/2212.11565). **[Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation](https://arxiv.org/abs/2212.11565)**
[Jay Zhangjie Wu](https://zhangjiewu.github.io/), [Yixiao Ge](https://geyixiao.com/), [Xintao Wang](https://xinntao.github.io/), [Stan Weixian Lei](), [Yuchao Gu](https://ycgu.site/), [Yufei Shi](), [Wynne Hsu](https://www.comp.nus.edu.sg/~whsu/), [Ying Shan](https://scholar.google.com/citations?user=4oXBp9UAAAAJ&hl=en), [Xiaohu Qie](https://scholar.google.com/citations?user=mk-F69UAAAAJ&hl=en), [Mike Zheng Shou](https://sites.google.com/view/showlab)
[![Project Website](https://img.shields.io/badge/Project-Website-orange)](https://tuneavideo.github.io/) [![arXiv](https://img.shields.io/badge/arXiv-2212.11565-b31b1b.svg)](https://arxiv.org/abs/2212.11565) [![Hugging Face Spaces](https://img.shields.io/badge/%F0%9F%A4%97%20Hugging%20Face-Spaces-blue)](https://huggingface.co/spaces/Tune-A-Video-library/Tune-A-Video-Training-UI) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/showlab/Tune-A-Video/blob/main/notebooks/Tune-A-Video.ipynb)


Given a video-text pair as input, our method, Tune-A-Video, fine-tunes a pre-trained text-to-image diffusion model for text-to-video generation.

## News - [02/22/2023] Improved consistency using DDIM inversion. - [02/08/2023] [Colab demo](https://colab.research.google.com/github/showlab/Tune-A-Video/blob/main/notebooks/Tune-A-Video.ipynb) released! - [02/03/2023] Pre-trained Tune-A-Video models are available on [Hugging Face Library](https://huggingface.co/Tune-A-Video-library)! - [01/28/2023] New Feature: tune a video on personalized [DreamBooth](https://dreambooth.github.io/) models. - [01/28/2023] Code released! ## Setup ### Requirements ```shell pip install -r requirements.txt ``` Installing [xformers](https://github.com/facebookresearch/xformers) is highly recommended for more efficiency and speed on GPUs. To enable xformers, set `enable_xformers_memory_efficient_attention=True` (default). ### Weights **[Stable Diffusion]** [Stable Diffusion](https://arxiv.org/abs/2112.10752) is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input. The pre-trained Stable Diffusion models can be downloaded from Hugging Face (e.g., [Stable Diffusion v1-4](https://huggingface.co/CompVis/stable-diffusion-v1-4), [v2-1](https://huggingface.co/stabilityai/stable-diffusion-2-1)). You can also use fine-tuned Stable Diffusion models trained on different styles (e.g, [Modern Disney](https://huggingface.co/nitrosocke/mo-di-diffusion), [Redshift](https://huggingface.co/nitrosocke/redshift-diffusion), etc.). **[DreamBooth]** [DreamBooth](https://dreambooth.github.io/) is a method to personalize text-to-image models like Stable Diffusion given just a few images (3~5 images) of a subject. Tuning a video on DreamBooth models allows personalized text-to-video generation of a specific subject. There are some public DreamBooth models available on [Hugging Face](https://huggingface.co/sd-dreambooth-library) (e.g., [mr-potato-head](https://huggingface.co/sd-dreambooth-library/mr-potato-head)). You can also train your own DreamBooth model following [this training example](https://github.com/huggingface/diffusers/tree/main/examples/dreambooth). ## Usage ### Training To fine-tune the text-to-image diffusion models for text-to-video generation, run this command: ```bash accelerate launch train_tuneavideo.py --config="configs/man-skiing.yaml" ``` Note: Tuning a 24-frame video usually takes `300~500` steps, about `10~15` minutes using one A100 GPU. Reduce `n_sample_frames` if your GPU memory is limited. ### Inference Once the training is done, run inference: ```python from tuneavideo.pipelines.pipeline_tuneavideo import TuneAVideoPipeline from tuneavideo.models.unet import UNet3DConditionModel from tuneavideo.util import save_videos_grid import torch pretrained_model_path = "./checkpoints/stable-diffusion-v1-4" my_model_path = "./outputs/man-skiing" unet = UNet3DConditionModel.from_pretrained(my_model_path, subfolder='unet', torch_dtype=torch.float16).to('cuda') pipe = TuneAVideoPipeline.from_pretrained(pretrained_model_path, unet=unet, torch_dtype=torch.float16).to("cuda") pipe.enable_xformers_memory_efficient_attention() pipe.enable_vae_slicing() prompt = "spider man is skiing" ddim_inv_latent = torch.load(f"{my_model_path}/inv_latents/ddim_latent-500.pt").to(torch.float16) video = pipe(prompt, latents=ddim_inv_latent, video_length=24, height=512, width=512, num_inference_steps=50, guidance_scale=12.5).videos save_videos_grid(video, f"./{prompt}.gif") ``` ## Results ### Pretrained T2I (Stable Diffusion)
Input Video Output Video
"A man is skiing" "Spider Man is skiing on the beach, cartoon styleā€¯ "Wonder Woman, wearing a cowboy hat, is skiing" "A man, wearing pink clothes, is skiing at sunset"
"A rabbit is eating a watermelon" "A rabbit is eating a watermelon on the table" "A cat with sunglasses is eating a watermelon on the beach" "A puppy is eating a cheeseburger on the table, comic style"
"A jeep car is moving on the road" "A Porsche car is moving on the beach" "A car is moving on the road, cartoon style" "A car is moving on the snow"
"A man is dribbling a basketball" "Trump is dribbling a basketball" "An astronaut is dribbling a basketball, cartoon style" "A lego man in a black suit is dribbling a basketball"
### Pretrained T2I (personalized DreamBooth)
Input Video Output Video
"A bear is playing guitar" "A rabbit is playing guitar, modern disney style" "A handsome prince is playing guitar, modern disney style" "A magic princess with sunglasses is playing guitar on the stage, modern disney style"
Input Video Output Video
"A bear is playing guitar" "Mr Potato Head, made of lego, is playing guitar on the snow" "Mr Potato Head, wearing sunglasses, is playing guitar on the beach" "Mr Potato Head is playing guitar in the starry night, Van Gogh style"
## Citation If you make use of our work, please cite our paper. ```bibtex @article{wu2022tuneavideo, title={Tune-A-Video: One-Shot Tuning of Image Diffusion Models for Text-to-Video Generation}, author={Wu, Jay Zhangjie and Ge, Yixiao and Wang, Xintao and Lei, Stan Weixian and Gu, Yuchao and Hsu, Wynne and Shan, Ying and Qie, Xiaohu and Shou, Mike Zheng}, journal={arXiv preprint arXiv:2212.11565}, year={2022} } ``` ## Shoutouts - This code builds on [diffusers](https://github.com/huggingface/diffusers). Thanks for open-sourcing! - Thanks [hysts](https://github.com/hysts) for the awesome [gradio demo](https://huggingface.co/spaces/Tune-A-Video-library/Tune-A-Video-Training-UI).