Improving Predictive State Representations via Gradient Descent

Authors

  • Nan Jiang University of Michigan
  • Alex Kulesza University of Michigan
  • Satinder Singh University of Michigan

DOI:

https://doi.org/10.1609/aaai.v30i1.10270

Keywords:

spectral learning, predictive state representation, gradient descent

Abstract

Predictive state representations (PSRs) model dynamical systems using appropriately chosen predictions about future observations as a representation of the current state. In contrast to the hidden states posited by HMMs or RNNs, PSR states are directly observable in the training data; this gives rise to a moment-matching spectral algorithm for learning PSRs that is computationally efficient and statistically consistent when the model complexity matches that of the true system generating the data. In practice, however, model mismatch is inevitable and while spectral learning remains appealingly fast and simple it may fail to find optimal models. To address this problem, we investigate the use of gradient methods for improving spectrally-learned PSRs. We show that only a small amount of additional gradient optimization can lead to significant performance gains, and moreover that initializing gradient methods with the spectral learning solution yields better models in significantly less time than starting from scratch.

Downloads

Published

2016-02-21

How to Cite

Jiang, N., Kulesza, A., & Singh, S. (2016). Improving Predictive State Representations via Gradient Descent. Proceedings of the AAAI Conference on Artificial Intelligence, 30(1). https://doi.org/10.1609/aaai.v30i1.10270

Issue

Section

Technical Papers: Machine Learning Methods