DocumentCode :
1483996
Title :
Reverberant Speech Segregation Based on Multipitch Tracking and Classification
Author :
Jin, Zhaozhang ; Wang, DeLiang
Author_Institution :
Dept. of Comput. Sci. & Eng., Ohio State Univ., Columbus, OH, USA
Volume :
19
Issue :
8
fYear :
2011
Firstpage :
2328
Lastpage :
2337
Abstract :
Room reverberation creates a major challenge to speech segregation. We propose a computational auditory scene analysis approach to monaural segregation of reverberant voiced speech, which performs multipitch tracking of reverberant mixtures and supervised classification. Speech and nonspeech models are separately trained, and each learns to map from a set of pitch-based features to a grouping cue which encodes the posterior probability of a time-frequency (T-F) unit being dominated by the source with the given pitch estimate. Because interference may be either speech or nonspeech, a likelihood ratio test selects the correct model for labeling corresponding T-F units. Experimental results show that the proposed system performs robustly in different types of interference and various reverberant conditions, and has a significant advantage over existing systems.
Keywords :
reverberation; speech processing; time-frequency analysis; computational auditory scene analysis; grouping cue; likelihood ratio test; monaural segregation; multipitch classification; multipitch tracking; reverberant speech segregation; reverberant voiced speech; room reverberation; Feature extraction; Harmonic analysis; Hidden Markov models; Interference; Labeling; Reverberation; Speech; Computational auditory scene analysis (CASA); monaural segregation; room reverberation; speech separation; supervised learning;
fLanguage :
English
Journal_Title :
Audio, Speech, and Language Processing, IEEE Transactions on
Publisher :
ieee
ISSN :
1558-7916
Type :
jour
DOI :
10.1109/TASL.2011.2134086
Filename :
5740581
Link To Document :
بازگشت