Title :
On-the-fly audio source separation
Author :
El Badawy, Dalia ; Duong, Ngoc Q. K. ; Ozerov, Alexey
Author_Institution :
Technicolor, Cesson-Sevigne, France
Abstract :
This paper addresses the challenging task of single channel audio source separation. We introduce a novel concept of on-the-fly audio source separation which greatly simplifies the user´s interaction with the system compared to the state-of-the-art user-guided approaches. In the proposed framework, the user is only asked to listen to an audio mixture and type some keywords (e.g. “dog barking”, “wind”, etc.) describing the sound sources to be separated. These keywords are then used as text queries to search for audio examples from the internet to guide the separation process. In particular, we propose several approaches to efficiently exploit these retrieved examples, including an approach based on a generic spectral model with group sparsity-inducing constraints. Finally, we demonstrate the effectiveness of the proposed framework with mixtures containing various types of sounds.
Keywords :
audio signal processing; source separation; audio mixture; on-the-fly audio source separation; separation process; sound sources; sparsity-inducing constraints; Abstracts; Engines; Matrix decomposition; On-the-fly source separation; group sparsity; non-negative matrix factorization; universal spectral model; user-guided;
Conference_Titel :
Machine Learning for Signal Processing (MLSP), 2014 IEEE International Workshop on
Conference_Location :
Reims
DOI :
10.1109/MLSP.2014.6958922