DocumentCode :
179600
Title :
GMM-free DNN acoustic model training
Author :
Senior, Alan ; Heigold, Georg ; Bacchiani, Michiel ; Liao, Haitao
Author_Institution :
Google Inc., New York, NY, USA
fYear :
2014
fDate :
4-9 May 2014
Firstpage :
5602
Lastpage :
5606
Abstract :
While deep neural networks (DNNs) have become the dominant acoustic model (AM) for speech recognition systems, they are still dependent on Gaussian mixture models (GMMs) for alignments both for supervised training and for context dependent (CD) tree building. Here we explore bootstrapping DNN AM training without GMM AMs and show that CD trees can be built with DNN alignments which are better matched to the DNN model and its features. We show that these trees and alignments result in better models than from the GMM alignments and trees. By removing the GMM acoustic model altogether we simplify the system required to train a DNN from scratch.
Keywords :
Gaussian processes; learning (artificial intelligence); mixture models; neural nets; speech recognition; statistical analysis; trees (mathematics); CD tree building; GMM-free DNN acoustic model training; Gaussian mixture models; bootstrapping DNN AM training; context dependent tree building; deep neural networks; speech recognition systems; supervised training; Acoustics; Context; Context modeling; Hidden Markov models; Neural networks; Speech recognition; Training; Deep neural networks; Viterbi forced-alignment; Voice Search; context dependent tree-building; flat start; hybrid neural network speech recognition; mobile speech recognition;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on
Conference_Location :
Florence
Type :
conf
DOI :
10.1109/ICASSP.2014.6854675
Filename :
6854675
Link To Document :
بازگشت