DGCN: Dynamic Graph Convolutional Network for Efficient Multi-Person Pose Estimation


  • Zhongwei Qiu University of Science and Technology Beijing
  • Kai Qiu Microsoft Research Asia
  • Jianlong Fu Microsoft Research Asia
  • Dongmei Fu University of Science and Technology Beijing




Multi-person pose estimation aims to detect human keypoints from images with multiple persons. Bottom-up methods for multi-person pose estimation have attracted extensive attention, owing to the good balance between efficiency and accuracy. Recent bottom-up methods usually follow the principle of keypoints localization and grouping, where relations between keypoints are the keys to group keypoints. These relations spontaneously construct a graph of keypoints, where the edges represent the relations between two nodes (i.e., keypoints). Existing bottom-up methods mainly define relations by empirically picking out edges from this graph, while omitting edges that may contain useful semantic relations. In this paper, we propose a novel Dynamic Graph Convolutional Module (DGCM) to model rich relations in the keypoints graph. Specifically, we take into account all relations (all edges of the graph) and construct dynamic graphs to tolerate large variations of human pose. The DGCM is quite lightweight, which allows it to be stacked like a pyramid architecture and learn structural relations from multi-level features. Our network with single DGCM based on ResNet-50 achieves relative gains of 3.2% and 4.8% over state-of-the-art bottom-up methods on COCO keypoints and MPII dataset, respectively.




How to Cite

Qiu, Z., Qiu, K., Fu, J., & Fu, D. (2020). DGCN: Dynamic Graph Convolutional Network for Efficient Multi-Person Pose Estimation. Proceedings of the AAAI Conference on Artificial Intelligence, 34(07), 11924-11931. https://doi.org/10.1609/aaai.v34i07.6867



AAAI Technical Track: Vision