Open-Loop Planning in Large-Scale Stochastic Domains

Authors

  • Ari Weinstein Rutgers University
  • Michael Littman Brown Univesity and Rutgers University

DOI:

https://doi.org/10.1609/aaai.v27i1.8547

Keywords:

reinforcement learning, receding horizon control, continuous Markov decision processes

Abstract

We focus on effective sample-based planning in the face of underactuation, high-dimensionality, drift, discrete system changes, and stochasticity. These are hallmark challenges for important problems, such as humanoid locomotion. In order to ensure broad applicability, we assume domain expertise is minimal and limited to a generative model. In order to make the method responsive, computational costs that scale linearly with the amount of samples taken from the generative model are required. We bring to bear a concrete method that satisfies all these requirements; it is a receding-horizon open-loop planner that employs cross-entropy optimization for policy construction. In simulation, we empirically demonstrate near-optimal decisions in a small domain and effective locomotion in several challenging humanoid control tasks.

Downloads

Published

2013-06-29

How to Cite

Weinstein, A., & Littman, M. (2013). Open-Loop Planning in Large-Scale Stochastic Domains. Proceedings of the AAAI Conference on Artificial Intelligence, 27(1), 1436-1442. https://doi.org/10.1609/aaai.v27i1.8547