Sound events often occur in unstructured environments where they exhibit wide variations in their frequency content and temporal structure. Convolutional neural networks (CNNs) are able to extract higher level features that are invariant to local spectral and temporal variations. Recurrent neural networks (RNNs) are powerful in learning the longer term temporal context in the audio signals. CNNs and RNNs as classifiers have recently shown improved performances over established methods in various sound recognition tasks. We combine these two approaches in a convolutional recurrent neural network (CRNN) and apply it on a polyphonic sound event detection task. We compare the performance of the proposed CRNN method with CNN, RNN, and other established methods, and observe a considerable improvement for four different datasets consisting of everyday sound events.
Convolutional Recurrent Neural Networks for Polyphonic Sound Event Detection
Emre Çakir,Giambattista Parascandolo,Toni Heittola,H. Huttunen,Tuomas Virtanen
Published 2017 in IEEE/ACM Transactions on Audio Speech and Language Processing
ABSTRACT
PUBLICATION RECORD
- Publication year
2017
- Venue
IEEE/ACM Transactions on Audio Speech and Language Processing
- Publication date
2017-02-21
- Fields of study
Computer Science
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-54 of 54 references · Page 1 of 1