Planning with Abstract Learned Models While Learning Transferable Subtasks

Authors

  • John Winder University of Maryland, Baltimore County
  • Stephanie Milani Carnegie Mellon University
  • Matthew Landen Georgia Institute of Technology
  • Erebus Oh University of Maryland, Baltimore County
  • Shane Parr University of Massachusetts Amherst
  • Shawn Squire University of Maryland, Baltimore County
  • Marie desJardins Simmons University
  • Cynthia Matuszek University of Maryland, Baltimore County

DOI:

https://doi.org/10.1609/aaai.v34i06.6555

Abstract

We introduce an algorithm for model-based hierarchical reinforcement learning to acquire self-contained transition and reward models suitable for probabilistic planning at multiple levels of abstraction. We call this framework Planning with Abstract Learned Models (PALM). By representing subtasks symbolically using a new formal structure, the lifted abstract Markov decision process (L-AMDP), PALM learns models that are independent and modular. Through our experiments, we show how PALM integrates planning and execution, facilitating a rapid and efficient learning of abstract, hierarchical models. We also demonstrate the increased potential for learned models to be transferred to new and related tasks.

Downloads

Published

2020-04-03

How to Cite

Winder, J., Milani, S., Landen, M., Oh, E., Parr, S., Squire, S., desJardins, M., & Matuszek, C. (2020). Planning with Abstract Learned Models While Learning Transferable Subtasks. Proceedings of the AAAI Conference on Artificial Intelligence, 34(06), 9992-10000. https://doi.org/10.1609/aaai.v34i06.6555

Issue

Section

AAAI Technical Track: Planning, Routing, and Scheduling