ReLUPruner: Rethinking ReLU Importance with Taylor Expansion for Efficient Private Inference

Authors

  • Zhenpeng Li Wuhan University
  • Jinshuo Liu Wuhan University
  • Xinyan Wang Wuhan University
  • Lina Wang Wuhan University
  • Jeff Z. Pan University of Edinburgh

DOI:

https://doi.org/10.1609/aaai.v40i28.39502

Abstract

With the growing adoption of Machine-Learning-As-A-Service (MLaaS), Private Inference (PI) has emerged as a promising solution to address its security concerns through cryptographic techniques. However, nonlinear operations in neural networks account for most of the computational and communication overhead in PI. Existing studies mainly focus on optimizing and reducing the number of ReLU activations in neural networks, but traditional pruning methods may mistakenly remove ReLUs that are critical to maintaining model accuracy. To accurately evaluate the importance of ReLUs in the network, we propose ReLUPruner, a method that uses Taylor expansion to quantify the impact on loss before and after ReLU replacement. Furthermore, we establish a hierarchical importance metric to guide layer-wise ReLU budget allocation and adopt a progressive pruning strategy that dynamically adjust the pruning rate of each layer according to training progress. Extensive experiments on various models and datasets show that ReLUPruner achieves a good balance between ReLU budget and model accuracy, yielding improvements of 1.89% (12.9k ReLUs, CIFAR-10), 3.62% (50k ReLUs, CIFAR-100) and 2.66% (30k ReLUs, Tiny-ImageNet) over the previous state-of-the-art.

Published

2026-03-14

How to Cite

Li, Z., Liu, J., Wang, X., Wang, L., & Pan, J. Z. (2026). ReLUPruner: Rethinking ReLU Importance with Taylor Expansion for Efficient Private Inference. Proceedings of the AAAI Conference on Artificial Intelligence, 40(28), 23328–23336. https://doi.org/10.1609/aaai.v40i28.39502

Issue

Section

AAAI Technical Track on Machine Learning V