HCPO: Hierarchical Conductor-Based Policy Optimization in Multi-Agent Reinforcement Learning
DOI:
https://doi.org/10.1609/aaai.v40i35.40199Abstract
In cooperative Multi-Agent Reinforcement Learning (MARL), efficient exploration is crucial for optimizing the performance of joint policy. However, existing methods often update joint policies via independent agent exploration, without coordination among agents, which inherently constrains the expressive capacity and exploration of joint policies. To address this issue, we propose a conductor-based joint policy framework that directly enhances the expressive capacity of joint policies and coordinates exploration. In addition, we develop a Hierarchical Conductor-based Policy Optimization (HCPO) algorithm that instructs policy updates for the conductor and agents in a direction aligned with performance improvement. A rigorous theoretical guarantee further establishes the monotonicity of the joint policy optimization process. By deploying local conductors, HCPO retains centralized training benefits while eliminating inter-agent communication during execution. Finally, we evaluate HCPO on three challenging benchmarks: StarCraft II Multi-agent Challenge, Multi-agent MuJoCo, and Multi-agent Particle Environment. The results indicate that HCPO outperforms competitive MARL baselines regarding cooperative efficiency and stability.Published
2026-03-14
How to Cite
Liu, Z., Tu, J., Hong, Y., Xiong, L., Jin, Y., Tang, Y., & Li, F. (2026). HCPO: Hierarchical Conductor-Based Policy Optimization in Multi-Agent Reinforcement Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 40(35), 29564-29572. https://doi.org/10.1609/aaai.v40i35.40199
Issue
Section
AAAI Technical Track on Multiagent Systems