GCNext: Towards the Unity of Graph Convolutions for Human Motion Prediction
DOI:
https://doi.org/10.1609/aaai.v38i6.28375Keywords:
CV: Video Understanding & Activity Analysis, CV: Applications, CV: Biometrics, Face, Gesture & Pose, CV: Motion & Tracking, CV: Scene Analysis & Understanding, CV: Vision for Robotics & Autonomous DrivingAbstract
The past few years has witnessed the dominance of Graph Convolutional Networks (GCNs) over human motion prediction. Various styles of graph convolutions have been proposed, with each one meticulously designed and incorporated into a carefully-crafted network architecture. This paper breaks the limits of existing knowledge by proposing Universal Graph Convolution (UniGC), a novel graph convolution concept that re-conceptualizes different graph convolutions as its special cases. Leveraging UniGC on network-level, we propose GCNext, a novel GCN-building paradigm that dynamically determines the best-fitting graph convolutions both sample-wise and layer-wise. GCNext offers multiple use cases, including training a new GCN from scratch or refining a preexisting GCN. Experiments on Human3.6M, AMASS, and 3DPW datasets show that, by incorporating unique module-to-network designs, GCNext yields up to 9x lower computational cost than existing GCN methods, on top of achieving state-of-the-art performance. Our code is available at https://github.com/BradleyWang0416/GCNext.Downloads
Published
2024-03-24
How to Cite
Wang, X., Cui, Q., Chen, C., & Liu, M. (2024). GCNext: Towards the Unity of Graph Convolutions for Human Motion Prediction. Proceedings of the AAAI Conference on Artificial Intelligence, 38(6), 5642-5650. https://doi.org/10.1609/aaai.v38i6.28375
Issue
Section
AAAI Technical Track on Computer Vision V