HardF-SNN: Hardware-Friendly Quantization for Spiking Neural Networks with Efficient Integer-Arithmetic-Only Inference

Authors

  • Hanwen Liu University of Electronic Science and Technology of China
  • Kexin Shi University of Electronic Science and Technology of China
  • Jieyuan Zhang University of Electronic Science and Technology of China
  • Yimeng Shan University of Electronic Science and Technology of China
  • Jibin Wu Hong Kong Polytechnic University
  • Wenyu Chen University of Electronic Science and Technology of China
  • Malu Zhang University of Electronic Science and Technology of China

DOI:

https://doi.org/10.1609/aaai.v40i3.37174

Abstract

Spiking Neural Networks (SNNs) are emerging as a promising energy-efficient alternative to Artificial Neural Networks (ANNs) due to their event-driven computation paradigm. However, recent advances toward large-scale high-performance SNNs inevitably lead to substantial memory and computational overhead. While quantization offers a potential way, many quantization approaches fail to deliver verifiable efficiency gains on resource-constrained hardware platforms. In this paper, we propose a lightweight and hardware-friendly SNN, termed HardF-SNN. Specifically, we first build a baseline model using shared-scale quantization and BN folding to simulate integer-only inference, as this has not been thoroughly discussed in prior SNN works. Then, through empirical and theoretical analysis, we identify that the baseline suffers from accuracy degradation and may cause training failure. To mitigate these issues, we propose proportional shared-scale quantization for enhanced dynamic range and integer-only BN using bit-shifting to stabilize training. Extensive experiments show that HardF-SNN achieves an optimal balance between performance and efficiency with excellent hardware compatibility. To demonstrate its effectiveness on resource-limited platforms, HardF-SNN is deployed on a dedicated FPGA-based hardware accelerator. Evaluation results indicate that our implementation achieves significant performance improvements over several existing hardware accelerators.

Downloads

Published

2026-03-14

How to Cite

Liu, H., Shi, K., Zhang, J., Shan, Y., Wu, J., Chen, W., & Zhang, M. (2026). HardF-SNN: Hardware-Friendly Quantization for Spiking Neural Networks with Efficient Integer-Arithmetic-Only Inference. Proceedings of the AAAI Conference on Artificial Intelligence, 40(3), 1937-1945. https://doi.org/10.1609/aaai.v40i3.37174

Issue

Section

AAAI Technical Track on Cognitive Modeling & Cognitive Systems