Title :
Deep recurrent de-noising auto-encoder and blind de-reverberation for reverberated speech recognition
Author :
Weninger, Felix ; Watanabe, Shigetaka ; Tachioka, Yuuki ; Schuller, Bjorn
Author_Institution :
Mitsubishi Electr. Res. Labs. (MERL), Cambridge, MA, USA
Abstract :
This paper describes our joint efforts to provide robust automatic speech recognition (ASR) for reverberated environments, such as in hands-free human-machine interaction. We investigate blind feature space de-reverberation and deep recurrent de-noising auto-encoders (DAE) in an early fusion scheme. Results on the 2014 REVERB Challenge development set indicate that the DAE front-end provides complementary performance gains to multi-condition training, feature transformations, and model adaptation. The proposed ASR system achieves word error rates of 17.62 % and 36.6 % on simulated and real data, which is a significant improvement over the Challenge baseline (25.16 and 47.2 %).
Keywords :
feature extraction; recurrent neural nets; reverberation; signal denoising; speech codecs; speech recognition; ASR; automatic speech recognition; blind dereverberation; blind feature space dereverberation; deep recurrent denoising auto-encoder; feature transformations; fusion scheme; hands-free human-machine interaction; model adaptation; multi-condition training; reverberated environments; reverberated speech recognition; word error rates; Adaptation models; Noise reduction; Reverberation; Speech; Speech recognition; Training; De-reverberation; automatic speech recognition; feature enhancement; recurrent neural networks;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on
Conference_Location :
Florence
DOI :
10.1109/ICASSP.2014.6854478