Approximate Linear Programming for Constrained Partially Observable Markov Decision Processes

Authors

  • Pascal Poupart University of Waterloo
  • Aarti Malhotra University of Waterloo
  • Pei Pei University of Waterloo
  • Kee-Eung Kim Korean Advanced Institute of Science and Technology
  • Bongseok Goh Korean Advanced Institute of Science and Technology
  • Michael Bowling University of Alberta

DOI:

https://doi.org/10.1609/aaai.v29i1.9655

Keywords:

Constrained POMDPs, Approximate Linear Programming, Finite State Controller

Abstract

In many situations, it is desirable to optimize a sequence of decisions by maximizing a primary objective while respecting some constraints with respect to secondary objectives. Such problems can be naturally modeled as constrained partially observable Markov decision processes (CPOMDPs) when the environment is partially observable. In this work, we describe a technique based on approximate linear programming to optimize policies in CPOMDPs. The optimization is performed offline and produces a finite state controller with desirable performance guarantees. The approach outperforms a constrained version of point-based value iteration on a suite of benchmark problems.

Downloads

Published

2015-03-04

How to Cite

Poupart, P., Malhotra, A., Pei, P., Kim, K.-E., Goh, B., & Bowling, M. (2015). Approximate Linear Programming for Constrained Partially Observable Markov Decision Processes. Proceedings of the AAAI Conference on Artificial Intelligence, 29(1). https://doi.org/10.1609/aaai.v29i1.9655