Subspace-Aware Exploration for Sparse-Reward Multi-Agent Tasks

Authors

  • Pei Xu School of Artificial Intelligence, University of Chinese Academy of Sciences CRISE, Institute of Automation, Chinese Academy of Sciences
  • Junge Zhang CRISE, Institute of Automation, Chinese Academy of Sciences
  • Qiyue Yin CRISE, Institute of Automation, Chinese Academy of Sciences
  • Chao Yu School of Computer Science and Engineering, Sun Yat-sen University
  • Yaodong Yang Beijing Institute for General AI Institute for AI, Peking University
  • Kaiqi Huang School of Artificial Intelligence, University of Chinese Academy of Sciences CRISE, Institute of Automation, Chinese Academy of Sciences CAS, Center for Excellence in Brain Science and Intelligence Technology

DOI:

https://doi.org/10.1609/aaai.v37i10.26384

Keywords:

MAS: Multiagent Learning, ML: Reinforcement Learning Algorithms

Abstract

Exploration under sparse rewards is a key challenge for multi-agent reinforcement learning problems. One possible solution to this issue is to exploit inherent task structures for an acceleration of exploration. In this paper, we present a novel exploration approach, which encodes a special structural prior on the reward function into exploration, for sparse-reward multi-agent tasks. Specifically, a novel entropic exploration objective which encodes the structural prior is proposed to accelerate the discovery of rewards. By maximizing the lower bound of this objective, we then propose an algorithm with moderate computational cost, which can be applied to practical tasks. Under the sparse-reward setting, we show that the proposed algorithm significantly outperforms the state-of-the-art algorithms in the multiple-particle environment, the Google Research Football and StarCraft II micromanagement tasks. To the best of our knowledge, on some hard tasks (such as 27m_vs_30m}) which have relatively larger number of agents and need non-trivial strategies to defeat enemies, our method is the first to learn winning strategies under the sparse-reward setting.

Downloads

Published

2023-06-26

How to Cite

Xu, P., Zhang, J., Yin, Q., Yu, C., Yang, Y., & Huang, K. (2023). Subspace-Aware Exploration for Sparse-Reward Multi-Agent Tasks. Proceedings of the AAAI Conference on Artificial Intelligence, 37(10), 11717-11725. https://doi.org/10.1609/aaai.v37i10.26384

Issue

Section

AAAI Technical Track on Multiagent Systems