Locality Matters: A Scalable Value Decomposition Approach for Cooperative Multi-Agent Reinforcement Learning

Authors

  • Roy Zohar The Hebrew University of Jerusalem
  • Shie Mannor Technion - Israel Institute of Technology
  • Guy Tennenholtz Technion - Israel Institute of Technology

DOI:

https://doi.org/10.1609/aaai.v36i8.20915

Keywords:

Machine Learning (ML)

Abstract

Cooperative multi-agent reinforcement learning (MARL) faces significant scalability issues due to state and action spaces that are exponentially large in the number of agents. As environments grow in size, effective credit assignment becomes increasingly harder and often results in infeasible learning times. Still, in many real-world settings, there exist simplified underlying dynamics that can be leveraged for more scalable solutions. In this work, we exploit such locality structures effectively whilst maintaining global cooperation. We propose a novel, value-based multi-agent algorithm called LOMAQ, which incorporates local rewards in the Centralized Training Decentralized Execution paradigm. Additionally, we provide a direct reward decomposition method for finding these local rewards when only a global signal is provided. We test our method empirically, showing it scales well compared to other methods, significantly improving performance and convergence speed.

Downloads

Published

2022-06-28

How to Cite

Zohar, R., Mannor, S., & Tennenholtz, G. (2022). Locality Matters: A Scalable Value Decomposition Approach for Cooperative Multi-Agent Reinforcement Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 36(8), 9278-9285. https://doi.org/10.1609/aaai.v36i8.20915

Issue

Section

AAAI Technical Track on Machine Learning III