Abstract

In this paper, we are interested in the human pose estimation problem with a focus on learning reliable high-resolution representations. Most existing methods recover high-resolution representations from low-resolution representations produced by a high-to-low resolution network. Instead, our proposed network maintains high-resolution representations through the whole process. We start from a high-resolution subnetwork as the first stage, gradually add high-to-low resolution subnetworks one by one to form more stages, and connect the mutli-resolution subnetworks in parallel. We conduct repeated multi-scale fusions such that each of the high-to-low resolution representations receives information from other parallel representations over and over, leading to rich high-resolution representations. As a result, the predicted keypoint heatmap is potentially more accurate and spatially more precise. We empirically demonstrate the effectiveness of our network through the superior pose estimation results over two benchmark datasets: the COCO keypoint detection dataset and the MPII Human Pose dataset. In addition, we show the superiority of our network in pose tracking on the PoseTrack dataset. The code and models have been publicly available at https://github.com/leoxiaobin/deep-high-resolution-net.pytorch.

Keywords

SubnetworkComputer scienceBenchmark (surveying)Artificial intelligenceCode (set theory)PoseResolution (logic)Focus (optics)Deep learningRepresentation (politics)High resolutionProcess (computing)Pattern recognition (psychology)Machine learningRemote sensing

Affiliated Institutions

Related Publications

Publication Info

Year
2019
Type
article
Pages
5686-5696
Citations
5184
Access
Closed

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

5184
OpenAlex
769
Influential
4219
CrossRef

Cite This

Ke Sun, Bin Xiao, Dong Liu et al. (2019). Deep High-Resolution Representation Learning for Human Pose Estimation. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) , 5686-5696. https://doi.org/10.1109/cvpr.2019.00584

Identifiers

DOI
10.1109/cvpr.2019.00584
arXiv
1902.09212

Data Quality

Data completeness: 84%