Maximum Reconstruction Estimation for Generative Latent-Variable Models

Authors

  • Yong Cheng Tsinghua University
  • Yang Liu Tsinghua University
  • Wei Xu Tsinghua University

DOI:

https://doi.org/10.1609/aaai.v31i1.10971

Keywords:

maximum reconstruction estimation

Abstract

Generative latent-variable models are important for natural language processing due to their capability of providing compact representations of data. As conventional maximum likelihood estimation (MLE) is prone to focus on explaining irrelevant but common correlations in data, we apply maximum reconstruction estimation (MRE) to learning generative latent-variable models alternatively, which aims to find model parameters that maximize the probability of reconstructing the observed data. We develop tractable algorithms to directly learn hidden Markov models and IBM translation models using the MRE criterion, without the need to introduce a separate reconstruction model to facilitate efficient inference. Experiments on unsupervised part-of-speech induction and unsupervised word alignment show that our approach enables generative latent-variable models to better discover intended correlations in data and outperforms maximum likelihood estimators significantly.

Downloads

Published

2017-02-12

How to Cite

Cheng, Y., Liu, Y., & Xu, W. (2017). Maximum Reconstruction Estimation for Generative Latent-Variable Models. Proceedings of the AAAI Conference on Artificial Intelligence, 31(1). https://doi.org/10.1609/aaai.v31i1.10971