Title :
Comparing Distance Measures for Hidden Markov Models
Author :
Mohammad, Maruf ; Tranter, W.H.
Author_Institution :
Mobile & Portable Radio Res. Group, Virginia Tech, VA
fDate :
March 31 2005-April 2 2005
Abstract :
In this paper, several distance measures for hidden Markov models (HMMs) are compared. The most commonly used distance measure between two HMMs is Kullback-Leibler divergence (KLD). Since there is no closed form solution, Monte-Carlo method is usually applied to calculate the KLD. However, the computational complexity in Monte-Carlo estimation may be prohibitive in practical applications, which motivated researchers to propose new distance measures for HMMs. Numerical examples are presented comparing three such distance measures against the Monte-Carlo method. Results show that it is possible to approximate the KLD with a saving of hundreds of times in computational complexity
Keywords :
computational complexity; hidden Markov models; speech recognition; HMM; Kullback-Leibler divergence; computational complexity; hidden Markov models; speech recognition; Automatic speech recognition; Closed-form solution; Computational complexity; Digital signal processing; Hidden Markov models; Probability distribution; Signal processing algorithms; Speech recognition; Statistics; Vocabulary;
Conference_Titel :
SoutheastCon, 2006. Proceedings of the IEEE
Conference_Location :
Memphis, TN
Print_ISBN :
1-4244-0168-2
DOI :
10.1109/second.2006.1629360