Listening to the World Improves Speech Command Recognition


  • Brian McMahan R7 Speech Sciences, Inc.
  • Delip Rao R7 Speech Sciences, Inc.



deep learning, transfer learning, speech recognition


We study transfer learning in convolutional network architectures applied to the task of recognizing audio, such as environmental sound events and speech commands. Our key finding is that not only is it possible to transfer representations from an unrelated task like environmental sound classification to a voice-focused task like speech command recognition, but also that doing so improves accuracies significantly. We also investigate the effect of increased model capacity for transfer learning audio, by first validating known results from the field of Computer Vision of achieving better accuracies with increasingly deeper networks on two audio datasets: UrbanSound8k and Google Speech Commands. Then we propose a simple multiscale input representation using dilated convolutions and show that it is able to aggregate larger contexts and increase classification performance. Further, the models trained using a combination of transfer learning and multiscale input representations need only 50% of the training data to achieve similar accuracies as a freshly trained model with 100% of the training data. Finally, we demonstrate a positive interaction effect for the multiscale input and transfer learning, making a case for the joint application of the two techniques.




How to Cite

McMahan, B., & Rao, D. (2018). Listening to the World Improves Speech Command Recognition. Proceedings of the AAAI Conference on Artificial Intelligence, 32(1).