Your browser doesn't support javascript.
loading
Configuration-Invariant Sound Localization Technique Using Azimuth-Frequency Representation and Convolutional Neural Networks.
Chun, Chanjun; Jeon, Kwang Myung; Choi, Wooyeol.
  • Chun C; Infrastructure Research Center, Korea Institute of Civil Engineering and Building Technology, Goyang 10223, Korea.
  • Jeon KM; IntFlow Co., Ltd, Gwangju 61080, Korea..
  • Choi W; Department of Computer Engineering, Chosun University, Gwangju 61452, Korea.
Sensors (Basel) ; 20(13)2020 Jul 05.
Article en En | MEDLINE | ID: mdl-32635619
Deep neural networks (DNNs) have achieved significant advancements in speech processing, and numerous types of DNN architectures have been proposed in the field of sound localization. When a DNN model is deployed for sound localization, a fixed input size is required. This is generally determined by the number of microphones, the fast Fourier transform size, and the frame size. if the numbers or configurations of the microphones change, the DNN model should be retrained because the size of the input features changes. in this paper, we propose a configuration-invariant sound localization technique using the azimuth-frequency representation and convolutional neural networks (CNNs). the proposed CNN model receives the azimuth-frequency representation instead of time-frequency features as the input features. the proposed model was evaluated in different environments from the microphone configuration in which it was originally trained. for evaluation, single sound source is simulated using the image method. Through the evaluations, it was confirmed that the localization performance was superior to the conventional steered response power phase transform (SRP-PHAT) and multiple signal classification (MUSIC) methods.
Palabras clave

Texto completo: 1 Banco de datos: MEDLINE Idioma: En Año: 2020 Tipo del documento: Article

Texto completo: 1 Banco de datos: MEDLINE Idioma: En Año: 2020 Tipo del documento: Article