Sparse Deep Transfer Learning for Convolutional Neural Network

Authors

  • Jiaming Liu Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences
  • Yali Wang Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences
  • Yu Qiao Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences

DOI:

https://doi.org/10.1609/aaai.v31i1.10801

Keywords:

Convolutional Neural Network, Deep Model Compression, Transfer Learning

Abstract

Extensive studies have demonstrated that the representations of convolutional neural networks (CNN), which are learned from a large-scale data set in the source domain, can be effectively transferred to a new target domain. However, compared to the source domain, the target domain often has limited data in practice. In this case, overfitting may significantly depress transferability, due to the model redundancy of the intensive CNN structures. To deal with this difficulty, we propose a novel sparse deep transfer learning approach for CNN. There are three main contributions in this work. First, we introduce a Sparse-SourceNet to reduce the redundancy in the source domain. Second, we introduce a Hybrid-TransferNet to improve the generalization ability and the prediction accuracy of transfer learning, by taking advantage of both model sparsity and implicit knowledge. Third, we introduce a Sparse-TargetNet, where we prune our Hybrid-TransferNet to obtain a highly-compact, source-knowledge-integrated CNN in the target domain. To examine the effectiveness of our methods, we perform our sparse deep transfer learning approach on a number of benchmark transfer learning tasks. The results show that, compared to the standard fine-tuning approach, our proposed approach achieves a significant pruning rate on CNN while improves the accuracy of transfer learning.

Downloads

Published

2017-02-13

How to Cite

Liu, J., Wang, Y., & Qiao, Y. (2017). Sparse Deep Transfer Learning for Convolutional Neural Network. Proceedings of the AAAI Conference on Artificial Intelligence, 31(1). https://doi.org/10.1609/aaai.v31i1.10801