Planning in Factored Action Spaces with Symbolic Dynamic Programming

Authors

  • Aswin Raghavan Oregon State University
  • Saket Joshi Oregon State University
  • Alan Fern Oregon State University
  • Prasad Tadepalli Oregon State University
  • Roni Khardon Tufts University

DOI:

https://doi.org/10.1609/aaai.v26i1.8364

Keywords:

Markov Decision Processes, Decision Theoretic Planning

Abstract

We consider symbolic dynamic programming (SDP) for solving Markov Decision Processes (MDP) with factored state and action spaces, where both states and actions are described by sets of discrete variables. Prior work on SDP has considered only the case of factored states and ignored structure in the action space, causing them to scale poorly in terms of the number of action variables. Our main contribution is to present the first SDP-based planning algorithm for leveraging both state and action space structure in order to compute compactly represented value functions and policies. Since our new algorithm can potentially require more space than when action structure is ignored, our second contribution is to describe an approach for smoothly trading-off space versus time via recursive conditioning. Finally, our third contribution is to introduce a novel SDP approximation that often significantly reduces planning time with little loss in quality by exploiting action structure in weakly coupled MDPs. We present empirical results in three domains with factored action spaces that show that our algorithms scale much better with the number of action variables as compared to state-of-the-art SDP algorithms.

Downloads

Published

2021-09-20

How to Cite

Raghavan, A., Joshi, S., Fern, A., Tadepalli, P., & Khardon, R. (2021). Planning in Factored Action Spaces with Symbolic Dynamic Programming. Proceedings of the AAAI Conference on Artificial Intelligence, 26(1), 1802-1808. https://doi.org/10.1609/aaai.v26i1.8364

Issue

Section

Reasoning about Plans, Processes and Actions