Updates πŸ”₯πŸ”₯πŸ”₯

We have released the Gradio demo for Hybrid (Trajectory + Landmark) Controls HERE!

Introduction

This repo provides the inference Gradio demo for Trajectory Control of MOFA-Video.

Environment Setup

pip install -r requirements.txt

Download checkpoints

  1. Download the pretrained checkpoints of SVD_xt from huggingface to ./ckpts.

  2. Download the checkpint of MOFA-Adapter from huggingface to ./ckpts.

The final structure of checkpoints should be:

./ckpts/
|-- controlnet
|   |-- config.json
|   `-- diffusion_pytorch_model.safetensors
|-- stable-video-diffusion-img2vid-xt-1-1
|   |-- feature_extractor
|       |-- ...
|   |-- image_encoder
|       |-- ...
|   |-- scheduler
|       |-- ...
|   |-- unet
|       |-- ...
|   |-- vae
|       |-- ...
|   |-- svd_xt_1_1.safetensors
|   `-- model_index.json

Run Gradio Demo

python run_gradio.py

Please refer to the instructions on the gradio interface during the inference process.

Paper

arxiv.org/abs/2405.20222

Downloads last month
0
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Space using MyNiuuu/MOFA-Video-Traj 1