malvika2003's picture
Upload folder using huggingface_hub
db5855f verified
# Image to Video Generation with Stable Video Diffusion
[Stable Video Diffusion (SVD)](https://stability.ai/stable-video) Image-to-Video is a diffusion model that takes in a still image as a conditioning frame, and generates a video from it. In this tutorial we consider how to convert and run Stable Video Diffusion using OpenVINO.
We will use [stable-video-diffusion-img2video-xt](https://huggingface.co/stabilityai/stable-video-diffusion-img2vid-xt) model as example. Additionally, to speedup video generation process we apply [AnimateLCM](https://arxiv.org/abs/2402.00769) LoRA weights and run optimization with [NNCF](https://github.com/openvinotoolkit/nncf/).
![result](https://github.com/openvinotoolkit/openvino_notebooks/assets/29454499/ae8a77b2-b5c9-45c5-a103-6e46c686739f)
## Notebook Contents
This notebook demonstrates how to convert and run stable video diffusion using OpenVINO.
Notebook contains the following steps:
- Create PyTorch model pipeline using Diffusers
- Convert Stable Video Diffusion Pipeline models to OpenVINO
- Convert Image Encoder
- Convert U-Net
- Convert VAE Encoder and Decoder
- Create Stable Video Diffusion Pipeline with OpenVINO
- Optimize pipeline with [NNCF](https://github.com/openvinotoolkit/nncf/)
- Compare results of original and optimized pipelines
- Interactive Demo
## Installation Instructions
This is a self-contained example that relies solely on its own code.</br>
We recommend running the notebook in a virtual environment. You only need a Jupyter server to start.
For details, please refer to [Installation Guide](../../README.md).