Robust and Efficient Transfer Learning with Hidden Parameter Markov Decision Processes

Authors

  • Taylor Killian Harvard University
  • George Konidaris Brown University
  • Finale Doshi-Velez Harvard University

DOI:

https://doi.org/10.1609/aaai.v31i1.11065

Keywords:

Reinforcement Learning, Transfer Learning, Latent Variable Models, Gaussian Process Dynamical Model

Abstract

An intriguing application of transfer learning emerges when tasks arise with similar, but not identical, dynamics. Hidden Parameter Markov Decision Processes (HiP-MDP) embed these tasks into a low-dimensional space; given the embedding parameters one can identify the MDP for a particular task. However, the original formulation of HiP-MDP had a critical flaw: the embedding uncertainty was modeled independently of the agent's state uncertainty, requiring an arduous training procedure. In this work, we apply a Gaussian Process latent variable model to jointly model the dynamics and the embedding, leading to a more elegant formulation, one that allows for better uncertainty quantification and thus more robust transfer.

Downloads

Published

2017-02-12

How to Cite

Killian, T., Konidaris, G., & Doshi-Velez, F. (2017). Robust and Efficient Transfer Learning with Hidden Parameter Markov Decision Processes. Proceedings of the AAAI Conference on Artificial Intelligence, 31(1). https://doi.org/10.1609/aaai.v31i1.11065