In a range of recent works, object-centric architectures have been shown to be suitable for unsupervised scene decomposition in the vision domain. Inspired by these methods we present AudioSlots, a slot-centric generative model for blind source separation in the audio domain. AudioSlots is built using permutation-equivariant encoder and decoder networks. The encoder network based on the Transformer architecture learns to map a mixed audio spectrogram to an unordered set of independent source embeddings. The spatial broadcast decoder network learns to generate the source spectrograms from the source embeddings. We train the model in an end-to-end manner using a permutation invariant loss function. Our results on Libri2Mix speech separation constitute a proof of concept that this approach shows promise. We discuss the results and limitations of our approach in detail, and further outline potential ways to overcome the limitations and directions for future work.
Audioslots: A Slot-Centric Generative Model For Audio Separation
P. Reddy,Scott Wisdom,Klaus Greff,J. Hershey,Thomas Kipf
Published 2023 in 2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)
ABSTRACT
PUBLICATION RECORD
- Publication year
2023
- Venue
2023 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)
- Publication date
2023-05-09
- Fields of study
Computer Science, Engineering
- Identifiers
- External record
- Source metadata
Semantic Scholar
CITATION MAP
EXTRACTION MAP
CLAIMS
- No claims are published for this paper.
CONCEPTS
- No concepts are published for this paper.
REFERENCES
Showing 1-38 of 38 references · Page 1 of 1
CITED BY
Showing 1-6 of 6 citing papers · Page 1 of 1