Iterative Project Quasi-Newton Algorithm for Training RBM

Authors

  • Shuai Mi Tianjin University
  • Xiaozhao Zhao Tianjin University
  • Yuexian Hou Tianjin University
  • Peng Zhang Tianjin University
  • Wenjie Li The Hong Kong Polytechnic University
  • Dawei Song Tianjin University

DOI:

https://doi.org/10.1609/aaai.v30i1.9936

Keywords:

RBM, Newton's method, Natural gradient

Abstract

The restricted Boltzmann machine (RBM) has been used as building blocks for many successful deep learning models, e.g., deep belief networks (DBN) and deep Boltzmann machine (DBM) etc. The training of RBM can be extremely slow in pathological regions. The second order optimization methods, such as quasi-Newton methods, were proposed to deal with this problem. However, the non-convexity results in many obstructions for training RBM, including the infeasibility of applying second order optimization methods. In order to overcome this obstruction, we introduce an em-like iterative project quasi-Newton (IPQN) algorithm. Specifically, we iteratively perform the sampling procedure where it is not necessary to update parameters, and the sub-training procedure that is convex. In sub-training procedures, we apply quasi-Newton methods to deal with the pathological problem. We further show that Newton's method turns out to be a good approximation of the natural gradient (NG) method in RBM training. We evaluate IPQN in a series of density estimation experiments on the artificial dataset and the MNIST digit dataset. Experimental results indicate that IPQN achieves an improved convergent performance over the traditional CD method.

Downloads

Published

2016-03-05

How to Cite

Mi, S., Zhao, X., Hou, Y., Zhang, P., Li, W., & Song, D. (2016). Iterative Project Quasi-Newton Algorithm for Training RBM. Proceedings of the AAAI Conference on Artificial Intelligence, 30(1). https://doi.org/10.1609/aaai.v30i1.9936