It is recommended to symlink the dataset root to $MMPOSE/data
.
If your folder structure is different, you may need to change the corresponding paths in config files.
MMPose supported datasets:
@article{h36m_pami,
author = {Ionescu, Catalin and Papava, Dragos and Olaru, Vlad and Sminchisescu, Cristian},
title = {Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments},
journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
publisher = {IEEE Computer Society},
volume = {36},
number = {7},
pages = {1325-1339},
month = {jul},
year = {2014}
}
For Human3.6M, please download from the official website and run the preprocessing script, which will extract camera parameters and pose annotations at full framerate (50 FPS) and downsampled framerate (10 FPS). The processed data should have the following structure:
mmpose
├── mmpose
├── docs
├── tests
├── tools
├── configs
`── data
├── h36m
├── annotation_body3d
| ├── cameras.pkl
| ├── fps50
| | ├── h36m_test.npz
| | ├── h36m_train.npz
| | ├── joint2d_rel_stats.pkl
| | ├── joint2d_stats.pkl
| | ├── joint3d_rel_stats.pkl
| | `── joint3d_stats.pkl
| `── fps10
| ├── h36m_test.npz
| ├── h36m_train.npz
| ├── joint2d_rel_stats.pkl
| ├── joint2d_stats.pkl
| ├── joint3d_rel_stats.pkl
| `── joint3d_stats.pkl
`── images
├── S1
| ├── S1_Directions_1.54138969
| | ├── S1_Directions_1.54138969_00001.jpg
| | ├── S1_Directions_1.54138969_00002.jpg
| | ├── ...
| ├── ...
├── S5
├── S6
├── S7
├── S8
├── S9
`── S11
@Article = {joo_iccv_2015,
author = {Hanbyul Joo, Hao Liu, Lei Tan, Lin Gui, Bart Nabbe, Iain Matthews, Takeo Kanade, Shohei Nobuhara, and Yaser Sheikh},
title = {Panoptic Studio: A Massively Multiview System for Social Motion Capture},
booktitle = {ICCV},
year = {2015}
}
Please follow voxelpose-pytorch to prepare this dataset.
Download the dataset by following the instructions in panoptic-toolbox and extract them under $MMPOSE/data/panoptic
.
Only download those sequences that are needed. You can also just download a subset of camera views by specifying the number of views (HD_Video_Number) and changing the camera order in ./scripts/getData.sh
. The used sequences and camera views can be found in VoxelPose. Note that the sequence "160906_band3" might not be available due to errors on the server of CMU Panoptic.
Note that we only use HD videos, calibration data, and 3D Body Keypoint in the codes. You can comment out other irrelevant codes such as downloading 3D Face data in ./scripts/getData.sh
.
The directory tree should be like this:
mmpose
├── mmpose
├── docs
├── tests
├── tools
├── configs
`── data
├── panoptic
├── 16060224_haggling1
| | ├── hdImgs
| | ├── hdvideos
| | ├── hdPose3d_stage1_coco19
| | ├── calibration_160224_haggling1.json
├── 160226_haggling1
├── ...
@inproceedings {belagian14multi,
title = {{3D} Pictorial Structures for Multiple Human Pose Estimation},
author = {Belagiannis, Vasileios and Amin, Sikandar and Andriluka, Mykhaylo and Schiele, Bernt and Navab
Nassir and Ilic, Slobo
booktitle = {IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2014},
month = {June},
organization={IEEE}
}
Please follow voxelpose-pytorch to prepare these two datasets.
Please download the datasets from the official website and extract them under $MMPOSE/data/campus
and $MMPOSE/data/shelf
, respectively. The original data include images as well as the ground truth pose file actorsGT.mat
.
We directly use the processed camera parameters from voxelpose-pytorch. You can download them from this repository and place in under $MMPOSE/data/campus/calibration_campus.json
and $MMPOSE/data/shelf/calibration_shelf.json
, respectively.
Like Voxelpose, due to the limited and incomplete annotations of the two datasets, we don't train the model using this dataset. Instead, we directly use the 2D pose estimator trained on COCO, and use independent 3D human poses from the CMU Panoptic dataset to train our 3D model. It lies in ${MMPOSE}/data/panoptic_training_pose.pkl
.
Like Voxelpose, for testing, we first estimate 2D poses and generate 2D heatmaps for these two datasets. You can download the predicted poses from voxelpose-pytorch and place them in $MMPOSE/data/campus/pred_campus_maskrcnn_hrnet_coco.pkl
and $MMPOSE/data/shelf/pred_shelf_maskrcnn_hrnet_coco.pkl
, respectively. You can also use the models trained on COCO dataset (like HigherHRNet) to generate 2D heatmaps directly.
The directory tree should be like this:
mmpose
├── mmpose
├── docs
├── tests
├── tools
├── configs
`── data
├── panoptic_training_pose.pkl
├── campus
| ├── Camera0
| | | ├── campus4-c0-00000.png
| | | ├── ...
| | | ├── campus4-c0-01999.png
| ...
| ├── Camera2
| | | ├── campus4-c2-00000.png
| | | ├── ...
| | | ├── campus4-c2-01999.png
| ├── calibration_campus.json
| ├── pred_campus_maskrcnn_hrnet_coco.pkl
| ├── actorsGT.mat
├── shelf
| ├── Camera0
| | | ├── img_000000.png
| | | ├── ...
| | | ├── img_003199.png
| ...
| ├── Camera4
| | | ├── img_000000.png
| | | ├── ...
| | | ├── img_003199.png
| ├── calibration_shelf.json
| ├── pred_shelf_maskrcnn_hrnet_coco.pkl
| ├── actorsGT.mat