Title :
Reverberant Speech Segregation Based on Multipitch Tracking and Classification
Author :
Jin, Zhaozhang ; Wang, DeLiang
Author_Institution :
Dept. of Comput. Sci. & Eng., Ohio State Univ., Columbus, OH, USA
Abstract :
Room reverberation creates a major challenge to speech segregation. We propose a computational auditory scene analysis approach to monaural segregation of reverberant voiced speech, which performs multipitch tracking of reverberant mixtures and supervised classification. Speech and nonspeech models are separately trained, and each learns to map from a set of pitch-based features to a grouping cue which encodes the posterior probability of a time-frequency (T-F) unit being dominated by the source with the given pitch estimate. Because interference may be either speech or nonspeech, a likelihood ratio test selects the correct model for labeling corresponding T-F units. Experimental results show that the proposed system performs robustly in different types of interference and various reverberant conditions, and has a significant advantage over existing systems.
Keywords :
reverberation; speech processing; time-frequency analysis; computational auditory scene analysis; grouping cue; likelihood ratio test; monaural segregation; multipitch classification; multipitch tracking; reverberant speech segregation; reverberant voiced speech; room reverberation; Feature extraction; Harmonic analysis; Hidden Markov models; Interference; Labeling; Reverberation; Speech; Computational auditory scene analysis (CASA); monaural segregation; room reverberation; speech separation; supervised learning;
Journal_Title :
Audio, Speech, and Language Processing, IEEE Transactions on
DOI :
10.1109/TASL.2011.2134086