Complex Dynamic Neurons Improved Spiking Transformer Network for Efficient Automatic Speech Recognition

Authors

  • Qingyu Wang Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences
  • Tielin Zhang Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences
  • Minglun Han Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences
  • Yi Wang School of Artificial Intelligence, Jilin University
  • Duzhen Zhang Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences
  • Bo Xu Institute of Automation, Chinese Academy of Sciences School of Artificial Intelligence, University of Chinese Academy of Sciences Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Sciences

DOI:

https://doi.org/10.1609/aaai.v37i1.25081

Keywords:

CMS: Brain Modeling, CMS: Agent & Cognitive Architectures, CMS: Simulating Humans, ML: Bio-Inspired Learning

Abstract

The spiking neural network (SNN) using leaky-integrated-and-fire (LIF) neurons has been commonly used in automatic speech recognition (ASR) tasks. However, the LIF neuron is still relatively simple compared to that in the biological brain. Further research on more types of neurons with different scales of neuronal dynamics is necessary. Here we introduce four types of neuronal dynamics to post-process the sequential patterns generated from the spiking transformer to get the complex dynamic neuron improved spiking transformer neural network (DyTr-SNN). We found that the DyTr-SNN could handle the non-toy automatic speech recognition task well, representing a lower phoneme error rate, lower computational cost, and higher robustness. These results indicate that the further cooperation of SNNs and neural dynamics at the neuron and network scales might have much in store for the future, especially on the ASR tasks.

Downloads

Published

2023-06-26

How to Cite

Wang, Q., Zhang, T., Han, M., Wang, Y., Zhang, D., & Xu, B. (2023). Complex Dynamic Neurons Improved Spiking Transformer Network for Efficient Automatic Speech Recognition. Proceedings of the AAAI Conference on Artificial Intelligence, 37(1), 102-109. https://doi.org/10.1609/aaai.v37i1.25081

Issue

Section

AAAI Technical Track on Cognitive Modeling & Cognitive Systems