REPrune: Channel Pruning via Kernel Representative Selection

Authors

  • Mincheol Park Yonsei University Korea Institute of Science and Technology
  • Dongjin Kim Korea University Korea Institute of Science and Technology
  • Cheonjun Park Yonsei University
  • Yuna Park Korea Institute of Science and Technology
  • Gyeong Eun Gong Hyundai MOBIS
  • Won Woo Ro Yonsei University
  • Suhyun Kim Korea Institute of Science and Technology

DOI:

https://doi.org/10.1609/aaai.v38i13.29370

Keywords:

ML: Learning on the Edge & Model Compression, ML: Auto ML and Hyperparameter Tuning

Abstract

Channel pruning is widely accepted to accelerate modern convolutional neural networks (CNNs). The resulting pruned model benefits from its immediate deployment on general-purpose software and hardware resources. However, its large pruning granularity, specifically at the unit of a convolution filter, often leads to undesirable accuracy drops due to the inflexibility of deciding how and where to introduce sparsity to the CNNs. In this paper, we propose REPrune, a novel channel pruning technique that emulates kernel pruning, fully exploiting the finer but structured granularity. REPrune identifies similar kernels within each channel using agglomerative clustering. Then, it selects filters that maximize the incorporation of kernel representatives while optimizing the maximum cluster coverage problem. By integrating with a simultaneous training-pruning paradigm, REPrune promotes efficient, progressive pruning throughout training CNNs, avoiding the conventional train-prune-finetune sequence. Experimental results highlight that REPrune performs better in computer vision tasks than existing methods, effectively achieving a balance between acceleration ratio and performance retention.

Published

2024-03-24

How to Cite

Park, M., Kim, D. ., Park, C., Park, Y., Gong, G. E., Ro, W. W., & Kim, S. (2024). REPrune: Channel Pruning via Kernel Representative Selection. Proceedings of the AAAI Conference on Artificial Intelligence, 38(13), 14545-14553. https://doi.org/10.1609/aaai.v38i13.29370

Issue

Section

AAAI Technical Track on Machine Learning IV