File size: 4,663 Bytes
d7a991a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
# 3D Body Keypoint Datasets

It is recommended to symlink the dataset root to `$MMPOSE/data`.
If your folder structure is different, you may need to change the corresponding paths in config files.

MMPose supported datasets:

- [Human3.6M](#human36m) \[ [Homepage](http://vision.imar.ro/human3.6m/description.php) \]
- [CMU Panoptic](#cmu-panoptic) \[ [Homepage](http://domedb.perception.cs.cmu.edu/) \]

## Human3.6M

<!-- [DATASET] -->

<details>
<summary align="right"><a href="https://ieeexplore.ieee.org/abstract/document/6682899/">Human3.6M (TPAMI'2014)</a></summary>

```bibtex
@article{h36m_pami,
  author = {Ionescu, Catalin and Papava, Dragos and Olaru, Vlad and Sminchisescu,  Cristian},
  title = {Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments},
  journal = {IEEE Transactions on Pattern Analysis and Machine Intelligence},
  publisher = {IEEE Computer Society},
  volume = {36},
  number = {7},
  pages = {1325-1339},
  month = {jul},
  year = {2014}
}
```

</details>

For [Human3.6M](http://vision.imar.ro/human3.6m/description.php), please download from the official website and run the [preprocessing script](/tools/dataset/preprocess_h36m.py), which will extract camera parameters and pose annotations at full framerate (50 FPS) and downsampled framerate (10 FPS). The processed data should have the following structure:

```text
mmpose
β”œβ”€β”€ mmpose
β”œβ”€β”€ docs
β”œβ”€β”€ tests
β”œβ”€β”€ tools
β”œβ”€β”€ configs
`── data
    β”œβ”€β”€ h36m
        β”œβ”€β”€ annotation_body3d
        |   β”œβ”€β”€ cameras.pkl
        |   β”œβ”€β”€ fps50
        |   |   β”œβ”€β”€ h36m_test.npz
        |   |   β”œβ”€β”€ h36m_train.npz
        |   |   β”œβ”€β”€ joint2d_rel_stats.pkl
        |   |   β”œβ”€β”€ joint2d_stats.pkl
        |   |   β”œβ”€β”€ joint3d_rel_stats.pkl
        |   |   `── joint3d_stats.pkl
        |   `── fps10
        |       β”œβ”€β”€ h36m_test.npz
        |       β”œβ”€β”€ h36m_train.npz
        |       β”œβ”€β”€ joint2d_rel_stats.pkl
        |       β”œβ”€β”€ joint2d_stats.pkl
        |       β”œβ”€β”€ joint3d_rel_stats.pkl
        |       `── joint3d_stats.pkl
        `── images
            β”œβ”€β”€ S1
            |   β”œβ”€β”€ S1_Directions_1.54138969
            |   |   β”œβ”€β”€ S1_Directions_1.54138969_00001.jpg
            |   |   β”œβ”€β”€ S1_Directions_1.54138969_00002.jpg
            |   |   β”œβ”€β”€ ...
            |   β”œβ”€β”€ ...
            β”œβ”€β”€ S5
            β”œβ”€β”€ S6
            β”œβ”€β”€ S7
            β”œβ”€β”€ S8
            β”œβ”€β”€ S9
            `── S11
```

Please note that Human3.6M dataset is also used in the [3D_body_mesh](/docs/en/tasks/3d_body_mesh.md) task, where different schemes for data preprocessing and organizing are adopted.

## CMU Panoptic

<details>
<summary align="right"><a href="https://openaccess.thecvf.com/content_iccv_2015/html/Joo_Panoptic_Studio_A_ICCV_2015_paper.html">CMU Panoptic (ICCV'2015)</a></summary>

```bibtex
@Article = {joo_iccv_2015,
author = {Hanbyul Joo, Hao Liu, Lei Tan, Lin Gui, Bart Nabbe, Iain Matthews, Takeo Kanade, Shohei Nobuhara, and Yaser Sheikh},
title = {Panoptic Studio: A Massively Multiview System for Social Motion Capture},
booktitle = {ICCV},
year = {2015}
}
```

</details>

Please follow [voxelpose-pytorch](https://github.com/microsoft/voxelpose-pytorch) to prepare this dataset.

1. Download the dataset by following the instructions in [panoptic-toolbox](https://github.com/CMU-Perceptual-Computing-Lab/panoptic-toolbox) and extract them under `$MMPOSE/data/panoptic`.

2. Only download those sequences that are needed. You can also just download a subset of camera views by specifying the number of views (HD_Video_Number) and changing the camera order in `./scripts/getData.sh`. The used sequences and camera views can be found in [VoxelPose](https://arxiv.org/abs/2004.06239). Note that the sequence "160906_band3" might not be available due to errors on the server of CMU Panoptic.

3. Note that we only use HD videos,  calibration data, and 3D Body Keypoint in the codes. You can comment out other irrelevant codes such as downloading 3D Face data in `./scripts/getData.sh`.

The directory tree should be like this:

```text
mmpose
β”œβ”€β”€ mmpose
β”œβ”€β”€ docs
β”œβ”€β”€ tests
β”œβ”€β”€ tools
β”œβ”€β”€ configs
`── data
    β”œβ”€β”€ panoptic
        β”œβ”€β”€ 16060224_haggling1
        |   |   β”œβ”€β”€ hdImgs
        |   |   β”œβ”€β”€ hdvideos
        |   |   β”œβ”€β”€ hdPose3d_stage1_coco19
        |   |   β”œβ”€β”€ calibration_160224_haggling1.json
        β”œβ”€β”€ 160226_haggling1
            β”œβ”€β”€ ...
```