DocumentCode :
730716
Title :
Context dependent phone models for LSTM RNN acoustic modelling
Author :
Senior, Andrew ; Sak, Hasim ; Shafran, Izhak
Author_Institution :
Google Inc., New York, NY, USA
fYear :
2015
fDate :
19-24 April 2015
Firstpage :
4585
Lastpage :
4589
Abstract :
Long Short Term Memory Recurrent Neural Networks (LSTM RNNs), combined with hidden Markov models (HMMs), have recently been show to outperform other acoustic models such as Gaussian mixture models (GMMs) and deep neural networks (DNNs) for large scale speech recognition. We argue that using multi-state HMMs with LSTM RNN acoustic models is an unnecessary vestige of GMM-HMM and DNN-HMM modelling since LSTM RNNs are able to predict output distributions through continuous, instead of piece-wise stationary, modelling of the acoustic trajectory. We demonstrate equivalent results for context independent whole-phone or 3-state models and show that minimum-duration modelling can lead to improved results. We go on to show that context dependent whole-phone models can perform as well as context dependent states, given a minimum duration model.
Keywords :
recurrent neural nets; speech recognition; LSTM RNN acoustic modelling; context dependent phone models; large scale speech recognition; long short term memory recurrent neural networks; multistate hidden Markov models; Acoustics; Context; Context modeling; Hidden Markov models; Recurrent neural networks; Speech recognition; Training; Hybrid neural networks; Long Short-Term Memory Recurrent Neural Networks; context dependent phone models; hidden Markov models;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location :
South Brisbane, QLD
Type :
conf
DOI :
10.1109/ICASSP.2015.7178839
Filename :
7178839
Link To Document :
بازگشت