Robust Policy Computation in Reward-Uncertain MDPs Using Nondominated Policies

Authors

  • Kevin Regan University of Toronto
  • Craig Boutilier University of Toronto

DOI:

https://doi.org/10.1609/aaai.v24i1.7740

Abstract

The precise specification of reward functions for Markov decision processes (MDPs) is often extremely difficult, motivating research into both reward elicitation and the robust solution of MDPs with imprecisely specified reward (IRMDPs). We develop new techniques for the robust optimization of IRMDPs, using the minimax regret decision criterion, that exploit the set of nondominated policies, i.e., policies that are optimal for some instantiation of the imprecise reward function. Drawing parallels to POMDP value functions, we devise a Witness-style algorithm for identifying nondominated policies. We also examine several new algorithms for computing minimax regret using the nondominated set, and examine both practically and theoretically the impact of approximating this set. Our results suggest that a small subset of the nondominated set can greatly speed up computation, yet yield very tight approximations to minimax regret.

Downloads

Published

2010-07-04

How to Cite

Regan, K., & Boutilier, C. (2010). Robust Policy Computation in Reward-Uncertain MDPs Using Nondominated Policies. Proceedings of the AAAI Conference on Artificial Intelligence, 24(1), 1127-1133. https://doi.org/10.1609/aaai.v24i1.7740

Issue

Section

Reasoning about Plans, Processes and Actions