Learning Methods to Generate Good Plans: Integrating HTN Learning and Reinforcement Learning

Authors

  • Chad Hogg Lehigh University
  • Ugur Kuter University of Maryland
  • Hector Munoz-Avila Lehigh University

DOI:

https://doi.org/10.1609/aaai.v24i1.7571

Keywords:

planning, hierarchical task networks, reinforcement learning

Abstract

We consider how to learn Hierarchical Task Networks (HTNs) for planning problems in which both the quality of solution plans generated by the HTNs and the speed at which those plans are found is important. We describe an integration of HTN Learning with Reinforcement Learning to both learn methods by analyzing semantic annotations on tasks and to produce estimates of the expected values of the learned methods by performing Monte Carlo updates. We performed an experiment in which plan quality was inversely related to plan length. In two planning domains, we evaluated the planning performance of the learned methods in comparison to two state-of-the-art satisficing classical planners, FastForward and SGPlan6, and one optimal planner, HSP*. The results demonstrate that a greedy HTN planner using the learned methods was able to generate higher quality solutions than SGPlan6 in both domains and FastForward in one. Our planner, FastForward, and SGPlan6 ran in similar time, while HSP* was exponentially slower.

Downloads

Published

2010-07-05

How to Cite

Hogg, C., Kuter, U., & Munoz-Avila, H. (2010). Learning Methods to Generate Good Plans: Integrating HTN Learning and Reinforcement Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 24(1), 1530-1535. https://doi.org/10.1609/aaai.v24i1.7571