DocumentCode :
179573
Title :
Recurrent deep neural networks for robust speech recognition
Author :
Chao Weng ; Dong Yu ; Watanabe, Shigetaka ; Juang, Biing-Hwang Fred
Author_Institution :
Georgia Inst. of Technol., Atlanta, GA, USA
fYear :
2014
fDate :
4-9 May 2014
Firstpage :
5532
Lastpage :
5536
Abstract :
In this work, we propose recurrent deep neural networks (DNNs) for robust automatic speech recognition (ASR). Full recurrent connections are added to certain hidden layer of a conventional feedforward DNN and allow the model to capture the temporal dependency in deep representations. A new backpropagation through time (BPTT) algorithm is introduced to make the minibatch stochastic gradient descent (SGD) on the proposed recurrent DNNs more efficient and effective. We evaluate the proposed recurrent DNN architecture under the hybrid setup on both the 2nd CHiME challenge (track 2) and Aurora-4 tasks. Experimental results on the CHiME challenge data show that the proposed system can obtain consistent 7% relative WER improvements over the DNN systems, achieving state-of-the-art performance without front-end preprocessing, speaker adaptive training or multiple decoding passes. For the experiments on Aurora-4, the proposed system achieves 4% relative WER improvement over a strong DNN baseline system.
Keywords :
backpropagation; feedforward neural nets; gradient methods; recurrent neural nets; speech recognition; ASR; BPTT algorithm; CHiME; SGD; backpropagation through time algorithm; feedforward DNN; minibatch stochastic gradient descent; recurrent DNN architecture; recurrent deep neural networks; robust automatic speech recognition; Neural networks; Noise; Robustness; Speech; Speech recognition; Training; Vectors; Aurora-4; CHiME; DNN; RNN; robust ASR;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on
Conference_Location :
Florence
Type :
conf
DOI :
10.1109/ICASSP.2014.6854661
Filename :
6854661
Link To Document :
بازگشت