DACOM: Learning Delay-Aware Communication for Multi-Agent Reinforcement Learning

Authors

  • Tingting Yuan University oft Göettingen
  • Hwei-Ming Chung University of Oslo NOOT Tech. Co., Ltd.
  • Jie Yuan Beijing University of Posts and Telecommunications
  • Xiaoming Fu University of Göettingen

DOI:

https://doi.org/10.1609/aaai.v37i10.26389

Keywords:

MAS: Agent Communication, ML: Reinforcement Learning Algorithms, ML: Reinforcement Learning Theory, MAS: Multiagent Learning

Abstract

Communication is supposed to improve multi-agent collaboration and overall performance in cooperative Multi-agent reinforcement learning (MARL). However, such improvements are prevalently limited in practice since most existing communication schemes ignore communication overheads (e.g., communication delays). In this paper, we demonstrate that ignoring communication delays has detrimental effects on collaborations, especially in delay-sensitive tasks such as autonomous driving. To mitigate this impact, we design a delay-aware multi-agent communication model (DACOM) to adapt communication to delays. Specifically, DACOM introduces a component, TimeNet, that is responsible for adjusting the waiting time of an agent to receive messages from other agents such that the uncertainty associated with delay can be addressed. Our experiments reveal that DACOM has a non-negligible performance improvement over other mechanisms by making a better trade-off between the benefits of communication and the costs of waiting for messages.

Downloads

Published

2023-06-26

How to Cite

Yuan, T., Chung, H.-M., Yuan, J., & Fu, X. (2023). DACOM: Learning Delay-Aware Communication for Multi-Agent Reinforcement Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 37(10), 11763-11771. https://doi.org/10.1609/aaai.v37i10.26389

Issue

Section

AAAI Technical Track on Multiagent Systems