Improved POMDP Tree Search Planning with Prioritized Action Branching
DOI:
https://doi.org/10.1609/aaai.v35i13.17412Keywords:
Planning with Markov Models (MDPs, POMDPs), Online Learning & Bandits, Planning under Uncertainty, Sequential Decision MakingAbstract
Online solvers for partially observable Markov decision processes have difficulty scaling to problems with large action spaces. This paper proposes a method called PA-POMCPOW to sample a subset of the action space that provides varying mixtures of exploitation and exploration for inclusion in a search tree. The proposed method first evaluates the action space according to a score function that is a linear combination of expected reward and expected information gain. The actions with the highest score are then added to the search tree during tree expansion. Experiments show that PA-POMCPOW is able to outperform existing state-of-the-art solvers on problems with large discrete action spaces.Downloads
Published
2021-05-18
How to Cite
Mern, J., Yildiz, A., Bush, L., Mukerji, T., & Kochenderfer, M. J. (2021). Improved POMDP Tree Search Planning with Prioritized Action Branching. Proceedings of the AAAI Conference on Artificial Intelligence, 35(13), 11888-11894. https://doi.org/10.1609/aaai.v35i13.17412
Issue
Section
AAAI Technical Track on Planning, Routing, and Scheduling