Decentralized Monte Carlo Tree Search for Partially Observable Multi-Agent Pathfinding
DOI:
https://doi.org/10.1609/aaai.v38i16.29703Keywords:
MAS: Multiagent Planning, ML: Reinforcement Learning, MAS: Multiagent Systems under UncertaintyAbstract
The Multi-Agent Pathfinding (MAPF) problem involves finding a set of conflict-free paths for a group of agents confined to a graph. In typical MAPF scenarios, the graph and the agents' starting and ending vertices are known beforehand, allowing the use of centralized planning algorithms. However, in this study, we focus on the decentralized MAPF setting, where the agents may observe the other agents only locally and are restricted in communications with each other. Specifically, we investigate the lifelong variant of MAPF, where new goals are continually assigned to the agents upon completion of previous ones. Drawing inspiration from the successful AlphaZero approach, we propose a decentralized multi-agent Monte Carlo Tree Search (MCTS) method for MAPF tasks. Our approach utilizes the agent's observations to recreate the intrinsic Markov decision process, which is then used for planning with a tailored for multi-agent tasks version of neural MCTS. The experimental results show that our approach outperforms state-of-the-art learnable MAPF solvers. The source code is available at https://github.com/AIRI-Institute/mats-lp.Downloads
Published
2024-03-24
How to Cite
Skrynnik, A., Andreychuk, A., Yakovlev, K., & Panov, A. (2024). Decentralized Monte Carlo Tree Search for Partially Observable Multi-Agent Pathfinding. Proceedings of the AAAI Conference on Artificial Intelligence, 38(16), 17531-17540. https://doi.org/10.1609/aaai.v38i16.29703
Issue
Section
AAAI Technical Track on Multiagent Systems