Spaces:
Build error
A newer version of the Gradio SDK is available:
5.29.0
2D Body Keypoint Datasets
It is recommended to symlink the dataset root to $MMPOSE/data
.
If your folder structure is different, you may need to change the corresponding paths in config files.
MMPose supported datasets:
- Images
- Videos
- PoseTrack18 [ Homepage ]
- sub-JHMDB [ Homepage ]
COCO
COCO (ECCV'2014)
@inproceedings{lin2014microsoft,
title={Microsoft coco: Common objects in context},
author={Lin, Tsung-Yi and Maire, Michael and Belongie, Serge and Hays, James and Perona, Pietro and Ramanan, Deva and Doll{\'a}r, Piotr and Zitnick, C Lawrence},
booktitle={European conference on computer vision},
pages={740--755},
year={2014},
organization={Springer}
}
For COCO data, please download from COCO download, 2017 Train/Val is needed for COCO keypoints training and validation. HRNet-Human-Pose-Estimation provides person detection result of COCO val2017 to reproduce our multi-person pose estimation results. Please download from OneDrive or GoogleDrive. Optionally, to evaluate on COCO'2017 test-dev, please download the image-info. Download and extract them under $MMPOSE/data, and make them look like this:
mmpose
βββ mmpose
βββ docs
βββ tests
βββ tools
βββ configs
`ββ data
βββ coco
β-- annotations
β β-- person_keypoints_train2017.json
β |-- person_keypoints_val2017.json
β |-- person_keypoints_test-dev-2017.json
|-- person_detection_results
| |-- COCO_val2017_detections_AP_H_56_person.json
| |-- COCO_test-dev2017_detections_AP_H_609_person.json
β-- train2017
β β-- 000000000009.jpg
β β-- 000000000025.jpg
β β-- 000000000030.jpg
β β-- ...
`-- val2017
β-- 000000000139.jpg
β-- 000000000285.jpg
β-- 000000000632.jpg
β-- ...
MPII
MPII (CVPR'2014)
@inproceedings{andriluka14cvpr,
author = {Mykhaylo Andriluka and Leonid Pishchulin and Peter Gehler and Schiele, Bernt},
title = {2D Human Pose Estimation: New Benchmark and State of the Art Analysis},
booktitle = {IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2014},
month = {June}
}
For MPII data, please download from MPII Human Pose Dataset. We have converted the original annotation files into json format, please download them from mpii_annotations. Extract them under {MMPose}/data, and make them look like this:
mmpose
βββ mmpose
βββ docs
βββ tests
βββ tools
βββ configs
`ββ data
βββ mpii
|ββ annotations
| |ββ mpii_gt_val.mat
| |ββ mpii_test.json
| |ββ mpii_train.json
| |ββ mpii_trainval.json
| `ββ mpii_val.json
`ββ images
|ββ 000001163.jpg
|ββ 000003072.jpg
During training and inference, the prediction result will be saved as '.mat' format by default. We also provide a tool to convert this '.mat' to more readable '.json' format.
python tools/dataset/mat2json ${PRED_MAT_FILE} ${GT_JSON_FILE} ${OUTPUT_PRED_JSON_FILE}
For example,
python tools/dataset/mat2json work_dirs/res50_mpii_256x256/pred.mat data/mpii/annotations/mpii_val.json pred.json
MPII-TRB
MPII-TRB (ICCV'2019)
@inproceedings{duan2019trb,
title={TRB: A Novel Triplet Representation for Understanding 2D Human Body},
author={Duan, Haodong and Lin, Kwan-Yee and Jin, Sheng and Liu, Wentao and Qian, Chen and Ouyang, Wanli},
booktitle={Proceedings of the IEEE International Conference on Computer Vision},
pages={9479--9488},
year={2019}
}
For MPII-TRB data, please download from MPII Human Pose Dataset. Please download the annotation files from mpii_trb_annotations. Extract them under {MMPose}/data, and make them look like this:
mmpose
βββ mmpose
βββ docs
βββ tests
βββ tools
βββ configs
`ββ data
βββ mpii
|ββ annotations
| |ββ mpii_trb_train.json
| |ββ mpii_trb_val.json
`ββ images
|ββ 000001163.jpg
|ββ 000003072.jpg
AIC
AI Challenger (ArXiv'2017)
@article{wu2017ai,
title={Ai challenger: A large-scale dataset for going deeper in image understanding},
author={Wu, Jiahong and Zheng, He and Zhao, Bo and Li, Yixin and Yan, Baoming and Liang, Rui and Wang, Wenjia and Zhou, Shipei and Lin, Guosen and Fu, Yanwei and others},
journal={arXiv preprint arXiv:1711.06475},
year={2017}
}
For AIC data, please download from AI Challenger 2017, 2017 Train/Val is needed for keypoints training and validation. Please download the annotation files from aic_annotations. Download and extract them under $MMPOSE/data, and make them look like this:
mmpose
βββ mmpose
βββ docs
βββ tests
βββ tools
βββ configs
`ββ data
βββ aic
β-- annotations
β β-- aic_train.json
β |-- aic_val.json
β-- ai_challenger_keypoint_train_20170902
β β-- keypoint_train_images_20170902
β β β-- 0000252aea98840a550dac9a78c476ecb9f47ffa.jpg
β β β-- 000050f770985ac9653198495ef9b5c82435d49c.jpg
β β β-- ...
`-- ai_challenger_keypoint_validation_20170911
β-- keypoint_validation_images_20170911
β-- 0002605c53fb92109a3f2de4fc3ce06425c3b61f.jpg
β-- 0003b55a2c991223e6d8b4b820045bd49507bf6d.jpg
β-- ...
CrowdPose
CrowdPose (CVPR'2019)
@article{li2018crowdpose,
title={CrowdPose: Efficient Crowded Scenes Pose Estimation and A New Benchmark},
author={Li, Jiefeng and Wang, Can and Zhu, Hao and Mao, Yihuan and Fang, Hao-Shu and Lu, Cewu},
journal={arXiv preprint arXiv:1812.00324},
year={2018}
}
For CrowdPose data, please download from CrowdPose. Please download the annotation files and human detection results from crowdpose_annotations. For top-down approaches, we follow CrowdPose to use the pre-trained weights of YOLOv3 to generate the detected human bounding boxes. For model training, we follow HigherHRNet to train models on CrowdPose train/val dataset, and evaluate models on CrowdPose test dataset. Download and extract them under $MMPOSE/data, and make them look like this:
mmpose
βββ mmpose
βββ docs
βββ tests
βββ tools
βββ configs
`ββ data
βββ crowdpose
β-- annotations
β β-- mmpose_crowdpose_train.json
β β-- mmpose_crowdpose_val.json
β β-- mmpose_crowdpose_trainval.json
β β-- mmpose_crowdpose_test.json
β β-- det_for_crowd_test_0.1_0.5.json
β-- images
β-- 100000.jpg
β-- 100001.jpg
β-- 100002.jpg
β-- ...
OCHuman
OCHuman (CVPR'2019)
@inproceedings{zhang2019pose2seg,
title={Pose2seg: Detection free human instance segmentation},
author={Zhang, Song-Hai and Li, Ruilong and Dong, Xin and Rosin, Paul and Cai, Zixi and Han, Xi and Yang, Dingcheng and Huang, Haozhi and Hu, Shi-Min},
booktitle={Proceedings of the IEEE conference on computer vision and pattern recognition},
pages={889--898},
year={2019}
}
For OCHuman data, please download the images and annotations from OCHuman, Move them under $MMPOSE/data, and make them look like this:
mmpose
βββ mmpose
βββ docs
βββ tests
βββ tools
βββ configs
`ββ data
βββ ochuman
β-- annotations
β β-- ochuman_coco_format_val_range_0.00_1.00.json
β |-- ochuman_coco_format_test_range_0.00_1.00.json
|-- images
β-- 000001.jpg
β-- 000002.jpg
β-- 000003.jpg
β-- ...
MHP
MHP (ACM MM'2018)
@inproceedings{zhao2018understanding,
title={Understanding humans in crowded scenes: Deep nested adversarial learning and a new benchmark for multi-human parsing},
author={Zhao, Jian and Li, Jianshu and Cheng, Yu and Sim, Terence and Yan, Shuicheng and Feng, Jiashi},
booktitle={Proceedings of the 26th ACM international conference on Multimedia},
pages={792--800},
year={2018}
}
For MHP data, please download from MHP. Please download the annotation files from mhp_annotations. Please download and extract them under $MMPOSE/data, and make them look like this:
mmpose
βββ mmpose
βββ docs
βββ tests
βββ tools
βββ configs
`ββ data
βββ mhp
β-- annotations
β β-- mhp_train.json
β β-- mhp_val.json
β
`-- train
β β-- images
β β β-- 1004.jpg
β β β-- 10050.jpg
β β β-- ...
β
`-- val
β β-- images
β β β-- 10059.jpg
β β β-- 10068.jpg
β β β-- ...
β
`-- test
β β-- images
β β β-- 1005.jpg
β β β-- 10052.jpg
β β β-- ...~~~~
PoseTrack18
PoseTrack18 (CVPR'2018)
@inproceedings{andriluka2018posetrack,
title={Posetrack: A benchmark for human pose estimation and tracking},
author={Andriluka, Mykhaylo and Iqbal, Umar and Insafutdinov, Eldar and Pishchulin, Leonid and Milan, Anton and Gall, Juergen and Schiele, Bernt},
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
pages={5167--5176},
year={2018}
}
For PoseTrack18 data, please download from PoseTrack18. Please download the annotation files from posetrack18_annotations. We have merged the video-wise separated official annotation files into two json files (posetrack18_train & posetrack18_val.json). We also generate the mask files to speed up training. For top-down approaches, we use MMDetection pre-trained Cascade R-CNN (X-101-64x4d-FPN) to generate the detected human bounding boxes. Please download and extract them under $MMPOSE/data, and make them look like this:
mmpose
βββ mmpose
βββ docs
βββ tests
βββ tools
βββ configs
`ββ data
βββ posetrack18
β-- annotations
β β-- posetrack18_train.json
β β-- posetrack18_val.json
β β-- posetrack18_val_human_detections.json
β β-- train
β β β-- 000001_bonn_train.json
β β β-- 000002_bonn_train.json
β β β-- ...
β β-- val
β β β-- 000342_mpii_test.json
β β β-- 000522_mpii_test.json
β β β-- ...
β `-- test
β β-- 000001_mpiinew_test.json
β β-- 000002_mpiinew_test.json
β β-- ...
β
`-- images
β β-- train
β β β-- 000001_bonn_train
β β β β-- 000000.jpg
β β β β-- 000001.jpg
β β β β-- ...
β β β-- ...
β β-- val
β β β-- 000342_mpii_test
β β β β-- 000000.jpg
β β β β-- 000001.jpg
β β β β-- ...
β β β-- ...
β `-- test
β β-- 000001_mpiinew_test
β β β-- 000000.jpg
β β β-- 000001.jpg
β β β-- ...
β β-- ...
`-- mask
β-- train
β β-- 000002_bonn_train
β β β-- 000000.jpg
β β β-- 000001.jpg
β β β-- ...
β β-- ...
`-- val
β-- 000522_mpii_test
β β-- 000000.jpg
β β-- 000001.jpg
β β-- ...
β-- ...
The official evaluation tool for PoseTrack should be installed from GitHub.
pip install git+https://github.com/svenkreiss/poseval.git
sub-JHMDB dataset
RSN (ECCV'2020)
@misc{cai2020learning,
title={Learning Delicate Local Representations for Multi-Person Pose Estimation},
author={Yuanhao Cai and Zhicheng Wang and Zhengxiong Luo and Binyi Yin and Angang Du and Haoqian Wang and Xinyu Zhou and Erjin Zhou and Xiangyu Zhang and Jian Sun},
year={2020},
eprint={2003.04030},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
For sub-JHMDB data, please download the images from JHMDB, Please download the annotation files from jhmdb_annotations. Move them under $MMPOSE/data, and make them look like this:
mmpose
βββ mmpose
βββ docs
βββ tests
βββ tools
βββ configs
`ββ data
βββ jhmdb
β-- annotations
β β-- Sub1_train.json
β |-- Sub1_test.json
β β-- Sub2_train.json
β |-- Sub2_test.json
β β-- Sub3_train.json
β |-- Sub3_test.json
|-- Rename_Images
β-- brush_hair
β β--April_09_brush_hair_u_nm_np1_ba_goo_0
| β β--00001.png
| β β--00002.png
β-- catch
β-- ...