ReX: An Efficient Approach to Reducing Memory Cost in Image Classification

Authors

  • Xuwei Qian Nanjing University of Information Science & Technology, Nanjing, China
  • Renlong Hang Nanjing University of Information Science & Technology, Nanjing, China
  • Qingshan Liu Nanjing University of Information Science & Technology, Nanjing, China

DOI:

https://doi.org/10.1609/aaai.v36i2.20106

Keywords:

Computer Vision (CV), Knowledge Representation And Reasoning (KRR)

Abstract

Exiting simple samples in adaptive multi-exit networks through early modules is an effective way to achieve high computational efficiency. One can observe that deployments of multi-exit architectures on resource-constrained devices are easily limited by high memory footprint of early modules. In this paper, we propose a novel approach named recurrent aggregation operator (ReX), which uses recurrent neural networks (RNNs) to effectively aggregate intra-patch features within a large receptive field to get delicate local representations, while bypassing large early activations. The resulting model, named ReXNet, can be easily extended to dynamic inference by introducing a novel consistency-based early exit criteria, which is based on the consistency of classification decisions over several modules, rather than the entropy of the prediction distribution. Extensive experiments on two benchmark datasets, i.e., Visual Wake Words, ImageNet-1k, demonstrate that our method consistently reduces the peak RAM and average latency of a wide variety of adaptive models on low-power devices.

Downloads

Published

2022-06-28

How to Cite

Qian, X., Hang, R., & Liu, Q. (2022). ReX: An Efficient Approach to Reducing Memory Cost in Image Classification. Proceedings of the AAAI Conference on Artificial Intelligence, 36(2), 2099-2107. https://doi.org/10.1609/aaai.v36i2.20106

Issue

Section

AAAI Technical Track on Computer Vision II