DocumentCode :
1394232
Title :
Comments on "Efficient training algorithms for HMMs using incremental estimation"
Author :
Byrne, William ; Gunawardana, Asela
Author_Institution :
Dept. of Electr. & Comput. Eng., Johns Hopkins Univ., Baltimore, MD, USA
Volume :
8
Issue :
6
fYear :
2000
Firstpage :
751
Lastpage :
754
Abstract :
The paper entitled "Efficient training algorithms for HMMs using incremental estimation" by Gotoh et al. (IEEE Trans. Speech Audio Processing, vol.6, p.539-48, Nov. 1998) investigated expectation maximization (EM) procedures that increase training speed. The claim of Gotoh et al. that these procedures are generalized EM (Dempster et al. 1977) procedures is shown to be incorrect in the present paper. We discuss why this is so, provide an example of nonmonotonic convergence to a local maximum in likelihood, and outline conditions that guarantee such convergence.
Keywords :
convergence of numerical methods; hidden Markov models; iterative methods; maximum likelihood estimation; speech processing; EM procedures; GEM; HMM; efficient training algorithms; expectation maximization algorithm; generalized EM methods; incremental estimation; local maximum likelihood; nonmonotonic convergence; training speed; Convergence; Hidden Markov models; Iterative algorithms; Iterative methods; Maximum likelihood estimation; Natural languages; Solids; Speech processing; Tomography; Training data;
fLanguage :
English
Journal_Title :
Speech and Audio Processing, IEEE Transactions on
Publisher :
ieee
ISSN :
1063-6676
Type :
jour
DOI :
10.1109/89.876315
Filename :
876315
Link To Document :
بازگشت