Compressed Conditional Mean Embeddings for Model-Based Reinforcement Learning

Authors

  • Guy Lever University College London
  • John Shawe-Taylor University College London
  • Ronnie Stafford University College London
  • Csaba Szepesvari University of Alberta

DOI:

https://doi.org/10.1609/aaai.v30i1.10304

Keywords:

Reinforcement Learning, Kernel Methods

Abstract

We present a model-based approach to solving Markov decision processes (MDPs) in which the system dynamics are learned using conditional mean embeddings (CMEs). This class of methods comes with strong performance guarantees, and enables planning to be performed in an induced finite (pseudo-)MDP, which approximates the MDP, but can be solved exactly using dynamic programming. Two drawbacks of existing methods exist: firstly, the size of the induced finite (pseudo-)MDP scales quadratically with the amount of data used to learn the model, costing much memory and time when planning with the learned model; secondly, learning the CME itself using powerful kernel least-squares is costly – a second computational bottleneck. We present an algorithm which maintains a rich kernelized CME model class, but solves both problems: firstly we demonstrate that the loss function for the CME model suggests a principled approach to compressing the induced (pseudo-)MDP, leading to faster planning, while maintaining guarantees; secondly we propose to learn the CME model itself using fast sparse-greedy kernel regression well-suited to the RL context. We demonstrate superior performance to existing methods in this class of modelbased approaches on a range of MDPs.

Downloads

Published

2016-02-21

How to Cite

Lever, G., Shawe-Taylor, J., Stafford, R., & Szepesvari, C. (2016). Compressed Conditional Mean Embeddings for Model-Based Reinforcement Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 30(1). https://doi.org/10.1609/aaai.v30i1.10304

Issue

Section

Technical Papers: Machine Learning Methods