Papers
arxiv:2109.11818

MODNet-V: Improving Portrait Video Matting via Background Restoration

Published on Sep 24, 2021
Authors:
,
,
,
,

Abstract

To address the challenging portrait video matting problem more precisely, existing works typically apply some matting priors that require additional user efforts to obtain, such as annotated trimaps or background images. In this work, we observe that instead of asking the user to explicitly provide a background image, we may recover it from the input video itself. To this end, we first propose a novel background restoration module (BRM) to recover the background image dynamically from the input video. BRM is extremely lightweight and can be easily integrated into existing matting models. By combining BRM with a recent image matting model, MODNet, we then present <PRE_TAG>MODNet-V</POST_TAG> for portrait video matting. Benefited from the strong background prior provided by BRM, <PRE_TAG>MODNet-V</POST_TAG> has only 1/3 of the parameters of MODNet but achieves comparable or even better performances. Our design allows <PRE_TAG>MODNet-V</POST_TAG> to be trained in an end-to-end manner on a single NVIDIA 3090 GPU. Finally, we introduce a new patch refinement module (PRM) to adapt <PRE_TAG>MODNet-V</POST_TAG> for high-resolution videos while keeping <PRE_TAG>MODNet-V</POST_TAG> lightweight and fast.

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2109.11818 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2109.11818 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2109.11818 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.