Your browser doesn't support javascript.
loading
Deep convolutional network for animal sound classification and source attribution using dual audio recordings.
Oikarinen, Tuomas; Srinivasan, Karthik; Meisner, Olivia; Hyman, Julia B; Parmar, Shivangi; Fanucci-Kiss, Adrian; Desimone, Robert; Landman, Rogier; Feng, Guoping.
Afiliação
  • Oikarinen T; McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, Massachusetts 02139, USA.
  • Srinivasan K; McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, Massachusetts 02139, USA.
  • Meisner O; McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, Massachusetts 02139, USA.
  • Hyman JB; McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, Massachusetts 02139, USA.
  • Parmar S; McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, Massachusetts 02139, USA.
  • Fanucci-Kiss A; McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, Massachusetts 02139, USA.
  • Desimone R; McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, Massachusetts 02139, USA.
  • Landman R; Stanley Center, Broad Institute, 57 Ames Street, Cambridge, Massachusetts 02139, USA.
  • Feng G; McGovern Institute for Brain Research, Massachusetts Institute of Technology, 43 Vassar Street, Cambridge, Massachusetts 02139, USA.
J Acoust Soc Am ; 145(2): 654, 2019 02.
Article em En | MEDLINE | ID: mdl-30823820
ABSTRACT
This paper introduces an end-to-end feedforward convolutional neural network that is able to reliably classify the source and type of animal calls in a noisy environment using two streams of audio data after being trained on a dataset of modest size and imperfect labels. The data consists of audio recordings from captive marmoset monkeys housed in pairs, with several other cages nearby. The network in this paper can classify both the call type and which animal made it with a single pass through a single network using raw spectrogram images as input. The network vastly increases data analysis capacity for researchers interested in studying marmoset vocalizations, and allows data collection in the home cage, in group housed animals.
Assuntos

Texto completo: 1 Bases de dados: MEDLINE Assunto principal: Vocalização Animal / Processamento de Sinais Assistido por Computador / Redes Neurais de Computação Limite: Animals Idioma: En Revista: J Acoust Soc Am Ano de publicação: 2019 Tipo de documento: Article País de afiliação: Estados Unidos

Texto completo: 1 Bases de dados: MEDLINE Assunto principal: Vocalização Animal / Processamento de Sinais Assistido por Computador / Redes Neurais de Computação Limite: Animals Idioma: En Revista: J Acoust Soc Am Ano de publicação: 2019 Tipo de documento: Article País de afiliação: Estados Unidos