DocumentCode
454734
Title
Gradient Boosting Learning of Hidden Markov Models
Author
Hu, Rusheng ; Li, Xiaolong ; Zhao, Yunxin
Author_Institution
Dept. of Comput. Sci., Missouri Univ., Columbia, MO
Volume
1
fYear
2006
fDate
14-19 May 2006
Abstract
In this paper, we present a new training algorithm, gradient boosting learning, for Gaussian mixture density (GMD) based acoustic models. This algorithm is based on a function approximation scheme from the perspective of optimization in function space rather than parameter space, i.e., stage-wise additive expansions of GMDs are used to search for optimal models instead of gradient descent optimization of model parameters. In the proposed approach, GMD starts from a single Gaussian and is built up by sequentially adding new components. Each new component is globally selected to produce optimal gain in the objective function. MLE and MMI are unified under the H-criterion, which is optimized by the extended BW (EBW) algorithm. A partial extended EM algorithm is developed for stage-wise optimization of new components. Experimental results on WSJ task demonstrate that the new algorithm leads to improved model quality and recognition performance
Keywords
Gaussian processes; acoustics; gradient methods; hidden Markov models; optimisation; speech recognition; Gaussian mixture density; acoustic models; function approximation scheme; gradient boosting learning; gradient descent optimization; hidden Markov models; Approximation algorithms; Boosting; Computer science; Error correction; Function approximation; Hidden Markov models; Maximum likelihood estimation; Mutual information; Speech recognition; Training data;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference on
Conference_Location
Toulouse
ISSN
1520-6149
Print_ISBN
1-4244-0469-X
Type
conf
DOI
10.1109/ICASSP.2006.1660233
Filename
1660233
Link To Document