VideoPainter: Any-length Video Inpainting and Editing with Plug-and-Play Context Control Paper • 2503.05639 • Published 6 days ago • 20
TrajectoryCrafter: Redirecting Camera Trajectory for Monocular Videos via Diffusion Models Paper • 2503.05638 • Published 6 days ago • 17
Layered Image Vectorization via Semantic Simplification Paper • 2406.05404 • Published Jun 8, 2024 • 3
NaturalL2S: End-to-End High-quality Multispeaker Lip-to-Speech Synthesis with Differential Digital Signal Processing Paper • 2502.12002 • Published 24 days ago • 1
RIFLEx: A Free Lunch for Length Extrapolation in Video Diffusion Transformers Paper • 2502.15894 • Published 20 days ago • 20
CAST: Component-Aligned 3D Scene Reconstruction from an RGB Image Paper • 2502.12894 • Published 23 days ago • 1
SongGen: A Single Stage Auto-regressive Transformer for Text-to-Song Generation Paper • 2502.13128 • Published 23 days ago • 37
TripoSG: High-Fidelity 3D Shape Synthesis using Large-Scale Rectified Flow Models Paper • 2502.06608 • Published Feb 10 • 33
Animate Anyone 2: High-Fidelity Character Image Animation with Environment Affordance Paper • 2502.06145 • Published Feb 10 • 16
Pippo: High-Resolution Multi-View Humans from a Single Image Paper • 2502.07785 • Published about 1 month ago • 11
Enhance-A-Video: Better Generated Video for Free Paper • 2502.07508 • Published about 1 month ago • 21
VidCRAFT3: Camera, Object, and Lighting Control for Image-to-Video Generation Paper • 2502.07531 • Published about 1 month ago • 13
FlashVideo:Flowing Fidelity to Detail for Efficient High-Resolution Video Generation Paper • 2502.05179 • Published Feb 7 • 24
Hibiki fr-en Collection Hibiki is a model for streaming speech translation , which can run on device! See https://github.com/kyutai-labs/hibiki. • 5 items • Updated Feb 6 • 50
OmniHuman-1: Rethinking the Scaling-Up of One-Stage Conditioned Human Animation Models Paper • 2502.01061 • Published Feb 3 • 189