Shakeout: A New Regularized Deep Neural Network Training Scheme

Authors

  • Guoliang Kang University of Technology Sydney
  • Jun Li University of Technology Sydney
  • Dacheng Tao University of Technology Sydney

DOI:

https://doi.org/10.1609/aaai.v30i1.10202

Abstract

Recent years have witnessed the success of deep neural networks in dealing with a plenty of practical problems. The invention of effective training techniques largely contributes to this success. The so-called "Dropout" training scheme is one of the most powerful tool to reduce over-fitting. From the statistic point of view, Dropout works by implicitly imposing an L2 regularizer on the weights. In this paper, we present a new training scheme: Shakeout. Instead of randomly discarding units as Dropout does at the training stage, our method randomly chooses to enhance or inverse the contributions of each unit to the next layer. We show that our scheme leads to a combination of L1 regularization and L2 regularization imposed on the weights, which has been proved effective by the Elastic Net models in practice.We have empirically evaluated the Shakeout scheme and demonstrated that sparse network weights are obtained via Shakeout training. Our classification experiments on real-life image datasets MNIST and CIFAR-10 show that Shakeout deals with over-fitting effectively.

Downloads

Published

2016-02-21

How to Cite

Kang, G., Li, J., & Tao, D. (2016). Shakeout: A New Regularized Deep Neural Network Training Scheme. Proceedings of the AAAI Conference on Artificial Intelligence, 30(1). https://doi.org/10.1609/aaai.v30i1.10202

Issue

Section

Technical Papers: Machine Learning Methods