--- language: - en - zh license: c-uda tags: - Video - Multi-viewpoint viewer: false --- # PKU-DyMVHumans Dataset ## Overview PKU-DyMVHumans is a versatile human-centric dataset designed for high-fidelity reconstruction and rendering of dynamic human performances in markerless multi-view capture settings. It comprises 32 humans across 45 different dynamic scenarios, each featuring highly detailed appearances and complex human motions. ### Sources - **Project page:** https://pku-dymvhumans.github.io - **Github:** https://github.com/zhengxyun/PKU-DyMVHumans - **Paper:** https://arxiv.org/abs/2403.16080 ### Key Features: - **High-fidelity performance:** We construct a multi-view system to capture humans in motion, containing 56/60 synchronous cameras with 1080P or 4K resolution. - **High-detailed appearance:** It captures complex cloth deformation, and intricate texture details, like delicate satin ribbon and special headwear. - **Complex human motion:** It covers a wide range of special costume performances, artistic movements, and sports activities. - **Human-object/scene interactions:** These include human-object interactions, as well as challenging multi-person interactions and complex scene effects (e.g., lighting, shadows, and smoking). ### Benchmark The objective of our benchmark is to achieve robust geometry reconstruction and novel view synthesis for dynamic humans under markerless and fixed multi-view camera settings, while minimizing the need for manual annotation and reducing time costs. This includes **neural scene decomposition**, **novel view synthesis**, and **dynamic human modeling**. ## Dataset Details ### Agreement Note that by downloading the datasets, you acknowledge that you have read the agreement, understand it, and agree to be bound by them: - The PKU-DyMVHumans dataset is made available only for non-commercial research purposes. Any other use, in particular any use for commercial purposes, is prohibited. - You agree not to further copy, publish or distribute any portion of the dataset. - Peking University reserves the right to terminate your access to the dataset at any time. ### Dataset Statistics - **Scenes:** 45 different dynamic scenarios, engaging in various actions and clothing styles. - **Actions:** 4 different action types: dance, kungfu, sport, and fashion show. - **Individual:** 32 professional players, including 16 males, 11 females, and 5 children. - **Frames:** totalling approximately 8.2 million frames. ## Dataset Structure For each scene, we provide the multi-view images (`./case_name/per_view/cam_*/images/`), the coarse foreground with RGBA channels (`./case_name/per_view/cam_*/images/`), as well as the coarse foreground segmentation (`./case_name/per_view/cam_*/pha/`), which are obtained using [BackgroundMattingV2](https://github.com/PeterL1n/BackgroundMattingV2). To make the benchmarks easier compare with our dataset, we save different data formats (i.e., [Surface-SOS](https://github.com/zhengxyun/Surface-SOS), [NeuS](https://github.com/Totoro97/NeuS), [NeuS2](https://github.com/19reborn/NeuS2), [Instant-ngp](https://github.com/NVlabs/instant-ngp), and [3D-Gaussian](https://github.com/graphdeco-inria/gaussian-splatting)) of PKU-DyMVHumans at **Part1** and write a document that describes the data process. ``` . |--- | |--- cams | |--- videos | |--- per_view | |--- per_frame | |--- data_ngp | |--- data_NeuS | |--- data_NeuS2 | |--- data_COLMAP | |--- |--- ... ``` ## BibTeX ``` @article{zheng2024DyMVHumans, title={PKU-DyMVHumans: A Multi-View Video Benchmark for High-Fidelity Dynamic Human Modeling}, author={Zheng, Xiaoyun and Liao, Liwei and Li, Xufeng and Jiao, Jianbo and Wang, Rongjie and Gao, Feng and Wang, Shiqi and Wang, Ronggang}, journal={IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}, year={2024} } ```