--- license: unknown --- # Stable Video Diffusion Temporal Controlnet ## Overview Introducing the Stable Video Diffusion Temporal Controlnet! This tool uses a controlnet style encoder with the svd base. It's designed to enhance your video diffusion projects by providing precise temporal control. ## Setup - **Controlnet Model:** download the inference repo from here: https://github.com/CiaraStrawberry/sdv_controlnet - **Installation:** run `pip install -r requirements.txt` - **Execution:** Run "run_inference.py". ## Demo ![combined_with_square_image_new_gif](https://github.com/CiaraStrawberry/sdv_controlnet/assets/13116982/055c8d3b-074e-4aeb-9ddc-70d12b5504d5) ## Notes - **Focus on Central Object:** The system tends to extract motion features primarily from a central object and, occasionally, from the background. It's best to avoid overly complex motion or obscure objects. - **Simplicity in Motion:** Stick to motions that svd can handle well without the controlnet. This ensures it will be able to apply the motion. ## Acknowledgements - **Diffusers Team:** For the svd implementation. - **Pixeli99:** For providing a practical svd training script: [SVD_Xtend](https://github.com/pixeli99/SVD_Xtend)