Relational Partially Observable MDPs

Authors

  • Chenggang Wang Tufts University
  • Roni Khardon Tufts University

DOI:

https://doi.org/10.1609/aaai.v24i1.7742

Abstract

Relational Markov Decision Processes (MDP) are a useful abstraction for stochastic planning problems since one can develop abstract solutions for them that are independent of domain size or instantiation. While there has been an increased interest in developing relational fully observable MDPs, there has been very little work on relational partially observable MDPs (POMDP), which deal with uncertainty in problem states in addition to stochastic action effects. This paper provides a concrete formalization of relational POMDPs making several technical contributions toward their solution. First, we show that to maintain correctness one must distinguish between quantification over states and quantification over belief states; this implies that solutions based on value iteration are inherently limited to the finite horizon case. Second, we provide a symbolic dynamic programing algorithm for finite horizon relational POMDPs, solving them at an abstract level, by lifting the propositional incremental pruning algorithm. Third, we show that this algorithm can be implemented using first order decision diagrams, a compact representation for functions over relational structures, that has been recently used to solve relational MDPs.

Downloads

Published

2010-07-04

How to Cite

Wang, C., & Khardon, R. (2010). Relational Partially Observable MDPs. Proceedings of the AAAI Conference on Artificial Intelligence, 24(1), 1153-1158. https://doi.org/10.1609/aaai.v24i1.7742

Issue

Section

Reasoning about Plans, Processes and Actions