HAVEN: Hierarchical Cooperative Multi-Agent Reinforcement Learning with Dual Coordination Mechanism
DOI:
https://doi.org/10.1609/aaai.v37i10.26386Keywords:
MAS: Multiagent LearningAbstract
Recently, some challenging tasks in multi-agent systems have been solved by some hierarchical reinforcement learning methods. Inspired by the intra-level and inter-level coordination in the human nervous system, we propose a novel value decomposition framework HAVEN based on hierarchical reinforcement learning for fully cooperative multi-agent problems. To address the instability arising from the concurrent optimization of policies between various levels and agents, we introduce the dual coordination mechanism of inter-level and inter-agent strategies by designing reward functions in a two-level hierarchy. HAVEN does not require domain knowledge and pre-training, and can be applied to any value decomposition variant. Our method achieves desirable results on different decentralized partially observable Markov decision process domains and outperforms other popular multi-agent hierarchical reinforcement learning algorithms.Downloads
Published
2023-06-26
How to Cite
Xu, Z., Bai, Y., Zhang, B., Li, D., & Fan, G. (2023). HAVEN: Hierarchical Cooperative Multi-Agent Reinforcement Learning with Dual Coordination Mechanism. Proceedings of the AAAI Conference on Artificial Intelligence, 37(10), 11735-11743. https://doi.org/10.1609/aaai.v37i10.26386
Issue
Section
AAAI Technical Track on Multiagent Systems