Coordinated Humanoid Robot Locomotion with Symmetry Equivariant Reinforcement Learning Policy

Authors

  • Buqing Nie MoE Key Lab of Artifcial Intelligence and AI Institute, Shanghai Jiao Tong University
  • Yang Zhang MoE Key Lab of Artifcial Intelligence and AI Institute, Shanghai Jiao Tong University
  • Rongjun Jin MoE Key Lab of Artifcial Intelligence and AI Institute, Shanghai Jiao Tong University
  • Zhanxiang Cao MoE Key Lab of Artifcial Intelligence and AI Institute, Shanghai Jiao Tong University Shanghai Innovation Institute
  • Huangxuan Lin MoE Key Lab of Artifcial Intelligence and AI Institute, Shanghai Jiao Tong University
  • Xiaokang Yang MoE Key Lab of Artifcial Intelligence and AI Institute, Shanghai Jiao Tong University
  • Yue Gao MoE Key Lab of Artifcial Intelligence and AI Institute, Shanghai Jiao Tong University Shanghai Innovation Institute

DOI:

https://doi.org/10.1609/aaai.v40i22.38918

Abstract

The human nervous system exhibits bilateral symmetry, enabling coordinated and balanced movements. However, existing Deep Reinforcement Learning (DRL) methods for humanoid robots neglect morphological symmetry of the robot, leading to uncoordinated and suboptimal behaviors. Inspired by human motor control, we propose Symmetry Equivariant Policy (SE-Policy), a new DRL framework that embeds strict symmetry equivariance in the actor and symmetry invariance in the critic without additional hyperparameters. SE-Policy enforces consistent behaviors across symmetric observations, producing temporally and spatially coordinated motions with higher task performance. Extensive experiments on velocity tracking tasks, conducted in both simulation and real-world deployment with the Unitree G1 humanoid robot, demonstrate that SE-Policy improves tracking accuracy by up to 40% compared to state-of-the-art baselines, while achieving superior spatial-temporal coordination. These results demonstrate the effectiveness of SE-Policy and its broad applicability to humanoid robots.

Downloads

Published

2026-03-14

How to Cite

Nie, B., Zhang, Y., Jin, R., Cao, Z., Lin, H., Yang, X., & Gao, Y. (2026). Coordinated Humanoid Robot Locomotion with Symmetry Equivariant Reinforcement Learning Policy. Proceedings of the AAAI Conference on Artificial Intelligence, 40(22), 18523–18531. https://doi.org/10.1609/aaai.v40i22.38918

Issue

Section

AAAI Technical Track on Intelligent Robotics