Optimizing Information Theory Based Bitwise Bottlenecks for Efficient Mixed-Precision Activation Quantization

Authors

  • Xichuan Zhou School of Microelectronics and Communication Engineering, Chongqing University Key Laboratory of Dependable Service Computing in Cyber-Physical-Society, Ministry of Education
  • Kui Liu School of Microelectronics and Communication Engineering, Chongqing University
  • Cong Shi School of Microelectronics and Communication Engineering, Chongqing University
  • Haijun Liu School of Microelectronics and Communication Engineering, Chongqing University
  • Ji Liu AI platform, Kwai Inc.

DOI:

https://doi.org/10.1609/aaai.v35i4.16474

Keywords:

Learning & Optimization for CV, Object Detection & Categorization, Vision for Robotics & Autonomous Driving, (Deep) Neural Network Algorithms

Abstract

Recent researches on information theory shed new light on the continuous attempts to open the black box of neural signal encoding. Inspired by the problem of lossy signal compression for wireless communication, this paper presents a Bitwise Bottleneck approach for quantizing and encoding neural network activations. Based on the rate-distortion theory, the Bitwise Bottleneck attempts to determine the most significant bits in activation representation by assigning and approximating the sparse coefficients associated with different bits. Given the constraint of a limited average code rate, the bottleneck minimizes the distortion for optimal activation quantization in a flexible layer-by-layer manner. Experiments over ImageNet and other datasets show that, by minimizing the quantization distortion of each layer, the neural network with bottlenecks achieves the state-of-the-art accuracy with low-precision activation. Meanwhile, by reducing the code rate, the proposed method can improve the memory and computational efficiency by over six times compared with the deep neural network with standard single-precision representation. The source code is available on GitHub: https://github.com/CQUlearningsystemgroup/BitwiseBottleneck.

Downloads

Published

2021-05-18

How to Cite

Zhou, X., Liu, K., Shi, C., Liu, H., & Liu, J. (2021). Optimizing Information Theory Based Bitwise Bottlenecks for Efficient Mixed-Precision Activation Quantization. Proceedings of the AAAI Conference on Artificial Intelligence, 35(4), 3590-3598. https://doi.org/10.1609/aaai.v35i4.16474

Issue

Section

AAAI Technical Track on Computer Vision III