Bayesian Unification of Sound Source Localization and Separation with Permutation Resolution

Authors

  • Takuma Otsuka Kyoto University
  • Katsuhiko Ishiguro NTT Corporation
  • Hiroshi Sawada NTT Corporation
  • Hiroshi Okuno Kyoto University

DOI:

https://doi.org/10.1609/aaai.v26i1.8376

Keywords:

Computational auditory scene analysis, Topic model, Variational Bayes inference

Abstract

Sound source localization and separation with permutation resolution are essential for achieving a computational auditory scene analysis system that can extract useful information from a mixture of various sounds. Because existing methods cope separately with these problems despite their mutual dependence, the overall result with these approaches can be degraded by any failure in one of these components. This paper presents a unified Bayesian framework to solve these problems simultaneously where localization and separation are regarded as a clustering problem. Experimental results confirm that our method outperforms state-of-the-art methods in terms of the separation quality with various setups including practical reverberant environments.

Downloads

Published

2021-09-20

How to Cite

Otsuka, T., Ishiguro, K., Sawada, H., & Okuno, H. (2021). Bayesian Unification of Sound Source Localization and Separation with Permutation Resolution. Proceedings of the AAAI Conference on Artificial Intelligence, 26(1), 2038-2045. https://doi.org/10.1609/aaai.v26i1.8376