Unsupervised Learning with Truncated Gaussian Graphical Models

Authors

  • Qinliang Su Duke University
  • Xuejun Liao Duke University
  • Chunyuan Li Duke University
  • Zhe Gan Duke University
  • Lawrence Carin Duke University

DOI:

https://doi.org/10.1609/aaai.v31i1.10815

Keywords:

Graphical Models, Latent variable model, Restricted Boltzmann Machine

Abstract

Gaussian graphical models (GGMs) are widely used for statistical modeling, because of ease of inference and the ubiquitous use of the normal distribution in practical approximations. However, they are also known for their limited modeling abilities, due to the Gaussian assumption. In this paper, we introduce a novel variant of GGMs, which relaxes the Gaussian restriction and yet admits efficient inference. Specifically, we impose a bipartite structure on the GGM and govern the hidden variables by truncated normal distributions. The nonlinearity of the model is revealed by its connection to rectified linear unit (ReLU) neural networks. Meanwhile, thanks to the bipartite structure and appealing properties of truncated normals, we are able to train the models efficiently using contrastive divergence. We consider three output constructs, accounting for real-valued, binary and count data. We further extend the model to deep constructions and show that deep models can be used for unsupervised pre-training of rectifier neural networks. Extensive experimental results are provided to validate the proposed models and demonstrate their superiority over competing models.

Downloads

Published

2017-02-13

How to Cite

Su, Q., Liao, X., Li, C., Gan, Z., & Carin, L. (2017). Unsupervised Learning with Truncated Gaussian Graphical Models. Proceedings of the AAAI Conference on Artificial Intelligence, 31(1). https://doi.org/10.1609/aaai.v31i1.10815