DocumentCode :
179245
Title :
Deep recurrent de-noising auto-encoder and blind de-reverberation for reverberated speech recognition
Author :
Weninger, Felix ; Watanabe, Shigetaka ; Tachioka, Yuuki ; Schuller, Bjorn
Author_Institution :
Mitsubishi Electr. Res. Labs. (MERL), Cambridge, MA, USA
fYear :
2014
fDate :
4-9 May 2014
Firstpage :
4623
Lastpage :
4627
Abstract :
This paper describes our joint efforts to provide robust automatic speech recognition (ASR) for reverberated environments, such as in hands-free human-machine interaction. We investigate blind feature space de-reverberation and deep recurrent de-noising auto-encoders (DAE) in an early fusion scheme. Results on the 2014 REVERB Challenge development set indicate that the DAE front-end provides complementary performance gains to multi-condition training, feature transformations, and model adaptation. The proposed ASR system achieves word error rates of 17.62 % and 36.6 % on simulated and real data, which is a significant improvement over the Challenge baseline (25.16 and 47.2 %).
Keywords :
feature extraction; recurrent neural nets; reverberation; signal denoising; speech codecs; speech recognition; ASR; automatic speech recognition; blind dereverberation; blind feature space dereverberation; deep recurrent denoising auto-encoder; feature transformations; fusion scheme; hands-free human-machine interaction; model adaptation; multi-condition training; reverberated environments; reverberated speech recognition; word error rates; Adaptation models; Noise reduction; Reverberation; Speech; Speech recognition; Training; De-reverberation; automatic speech recognition; feature enhancement; recurrent neural networks;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on
Conference_Location :
Florence
Type :
conf
DOI :
10.1109/ICASSP.2014.6854478
Filename :
6854478
Link To Document :
بازگشت