Fast and Accurate Binary Neural Networks Based on Depth-Width Reshaping

Authors

  • Ping Xue School of Computer Science and Information Engineering, Hefei University of Technology, Hefei, China
  • Yang Lu School of Computer Science and Information Engineering, Hefei University of Technology, Hefei, China Anhui Mine IOT and Security Monitoring Technology Key Laboratory, Hefei, China Engineering Research Center of Safety Critical Industrial Measurement and Control Technology, Ministry of Education, Hefei University of Technology, Hefei, China
  • Jingfei Chang School of Computer Science and Information Engineering, Hefei University of Technology, Hefei, China
  • Xing Wei School of Computer Science and Information Engineering, Hefei University of Technology, Hefei, China Anhui Mine IOT and Security Monitoring Technology Key Laboratory, Hefei, China Intelligent Manufacturing Institute of HeFei University of Technology, Hefei, China
  • Zhen Wei School of Computer Science and Information Engineering, Hefei University of Technology, Hefei, China Anhui Mine IOT and Security Monitoring Technology Key Laboratory, Hefei, China Engineering Research Center of Safety Critical Industrial Measurement and Control Technology, Ministry of Education, Hefei University of Technology, Hefei, China

DOI:

https://doi.org/10.1609/aaai.v37i9.26268

Keywords:

ML: Learning on the Edge & Model Compression

Abstract

Network binarization (i.e., binary neural networks, BNNs) can efficiently compress deep neural networks and accelerate model inference but cause severe accuracy degradation. Existing BNNs are mainly implemented based on the commonly used full-precision network backbones, and then the accuracy is improved with various techniques. However, there is a question of whether the full-precision network backbone is well adapted to BNNs. We start from the factors of the performance degradation of BNNs and analyze the problems of directly using full-precision network backbones for BNNs: for a given computational budget, the backbone of a BNN may need to be shallower and wider compared to the backbone of a full-precision network. With this in mind, Depth-Width Reshaping (DWR) is proposed to reshape the depth and width of existing full-precision network backbones and further optimize them by incorporating pruning techniques to better fit the BNNs. Extensive experiments demonstrate the analytical result and the effectiveness of the proposed method. Compared with the original backbones, the DWR backbones constructed by the proposed method result in close to O(√s) decrease in activations, while achieving an absolute accuracy increase by up to 1.7% with comparable computational cost. Besides, by using the DWR backbones, existing methods can achieve new state-of-the-art (SOTA) accuracy (e.g., 67.2% on ImageNet with ResNet-18 as the original backbone). We hope this work provides a novel insight into the backbone design of BNNs. The code is available at https://github.com/pingxue-hfut/DWR.

Downloads

Published

2023-06-26

How to Cite

Xue, P., Lu, Y., Chang, J., Wei, X., & Wei, Z. (2023). Fast and Accurate Binary Neural Networks Based on Depth-Width Reshaping. Proceedings of the AAAI Conference on Artificial Intelligence, 37(9), 10684-10692. https://doi.org/10.1609/aaai.v37i9.26268

Issue

Section

AAAI Technical Track on Machine Learning IV