Papers
arxiv:2411.02394

AutoVFX: Physically Realistic Video Editing from Natural Language Instructions

Published on Nov 4
· Submitted by haoyuhsu on Nov 5
Authors:
,
,
,

Abstract

Modern visual effects (VFX) software has made it possible for skilled artists to create imagery of virtually anything. However, the creation process remains laborious, complex, and largely inaccessible to everyday users. In this work, we present AutoVFX, a framework that automatically creates realistic and dynamic VFX videos from a single video and natural language instructions. By carefully integrating neural scene modeling, LLM-based code generation, and physical simulation, AutoVFX is able to provide physically-grounded, photorealistic editing effects that can be controlled directly using natural language instructions. We conduct extensive experiments to validate AutoVFX's efficacy across a diverse spectrum of videos and instructions. Quantitative and qualitative results suggest that AutoVFX outperforms all competing methods by a large margin in generative quality, instruction alignment, editing versatility, and physical plausibility.

Community

Paper author Paper submitter
edited 11 days ago
Paper author Paper submitter
This comment has been hidden
Paper author Paper submitter

We are thrilled to announce AutoVFX, a end-to-end system that allows physically realistic video editing based on textual editing instructions. Looking forward to any discussions here.

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2411.02394 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2411.02394 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2411.02394 in a Space README.md to link it from this page.

Collections including this paper 6