Bayes-Adaptive Interactive POMDPs

Authors

  • Brenda Ng Lawrence Livermore National Laboratory
  • Kofi Boakye Lawrence Livermore National Laboratory
  • Carol Meyers Lawrence Livermore National Laboratory
  • Andrew Wang Massachusetts Institute of Technology

DOI:

https://doi.org/10.1609/aaai.v26i1.8264

Keywords:

Multiagent Learning, Multiagent Planning, Reasoning under Uncertainty, Sequential Decision Making

Abstract

We introduce the Bayes-Adaptive Interactive Partially Observable Markov Decision Process (BA-IPOMDP), the first multiagent decision model that explicitly incorporates model learning. As in I-POMDPs, the BA-IPOMDP agent maintains beliefs over interactive states, which include the physical states as well as the other agents’ models. The BA-IPOMDP assumes that the state transition and observation probabilities are unknown, and augments the interactive states to include these parameters. Beliefs are maintained over this augmented interactive state space. This (necessary) state expansion exacerbates the curse of dimensionality, especially since each I-POMDP belief update is already a recursive procedure (because an agent invokes belief updates from other agents’ perspectives as part of its own belief update, in order to anticipate other agents’ actions). We extend the interactive particle filter to perform approximate belief update on BA-IPOMDPs. We present our findings on the multiagent Tiger problem.

Downloads

Published

2021-09-20

How to Cite

Ng, B., Boakye, K., Meyers, C., & Wang, A. (2021). Bayes-Adaptive Interactive POMDPs. Proceedings of the AAAI Conference on Artificial Intelligence, 26(1), 1408-1414. https://doi.org/10.1609/aaai.v26i1.8264

Issue

Section

AAAI Technical Track: Multiagent Systems