Training-Free ANN-to-SNN Conversion for High-Performance Spiking Transformers
DOI:
https://doi.org/10.1609/aaai.v40i3.37195Abstract
Leveraging the event-driven paradigm, Spiking Neural Networks (SNNs) offer a promising approach for constructing energy-efficient Transformer architectures. Compared to directly trained Spiking Transformers, ANN-to-SNN conversion methods bypass the high training costs. However, existing methods still suffer from notable limitations, failing to effectively handle nonlinear operations in Transformer architectures and requiring additional fine-tuning processes for pre-trained ANNs. To address these issues, we propose a high-performance and training-free ANN-to-SNN conversion framework tailored for Transformer architectures. Specifically, we introduce a Multi-basis Exponential Decay (MBE) neuron, which employs an exponential decay strategy and multi-basis encoding method to efficiently approximate various nonlinear operations. It removes the requirement for weight modifications in pre-trained ANNs. Extensive experiments across diverse tasks (CV, NLU, NLG) and mainstream Transformer architectures (ViT, RoBERTa, GPT-2) demonstrate that our method achieves near-lossless conversion accuracy with significantly lower latency. This provides a promising pathway for the efficient and scalable deployment of Spiking Transformers in real-world applications.Downloads
Published
2026-03-14
How to Cite
Wang, J., Deng, X., Wei, W., Zhang, D., Wang, S., Sun, Q., Zhang, J., Liu, H., Xie, N., & Zhang, M. (2026). Training-Free ANN-to-SNN Conversion for High-Performance Spiking Transformers. Proceedings of the AAAI Conference on Artificial Intelligence, 40(3), 2128-2136. https://doi.org/10.1609/aaai.v40i3.37195
Issue
Section
AAAI Technical Track on Cognitive Modeling & Cognitive Systems