DocumentCode :
336267
Title :
Structure and parameter learning via entropy minimization, with applications to mixture and hidden Markov models
Author :
Brand, Matthew
Author_Institution :
Mitsubishi Electr. Res. Lab., Cambridge, MA, USA
Volume :
3
fYear :
1999
fDate :
15-19 Mar 1999
Firstpage :
1749
Abstract :
We develop a computationally efficient framework for finding compact and highly accurate hidden-variable models via entropy minimization. The main results are: (1) an entropic prior that favors small, unambiguous, maximally structured models. (2) A prior balancing manipulation of Bayes´ rule that allows one to gradually introduce or remove constraints in the course of iterative reestimation. (1) and (2) combined give the information-theoretic free energy of the model and the means to manipulate it. (3) Maximum a posteriori (MAP) estimators such that entropy optimization and deterministic annealing can be performed wholly within expectation maximization (EM). (4) Trimming tests that identify excess parameters whose removal will increase the posterior, thereby simplifying the model and preventing over-fitting. The end result is a fast and exact hill-climbing algorithm that mixes continuous and combinatoric optimization and evades sub-optimal equilibria. Examples are given using speech and language problems
Keywords :
Bayes methods; computational linguistics; hidden Markov models; iterative methods; learning (artificial intelligence); maximum likelihood estimation; minimum entropy methods; optimisation; speech processing; Bayes rule; MAP estimator; combinatoric optimization; computationally efficient framework; constraints; continuous optimization; deterministic annealing; entropic prior; entropy minimization; entropy optimization; excess parameters; expectation maximization; hidden Markov models; hidden-variable models; hill-climbing algorithm; information-theoretic free energy; iterative reestimation; maximum a posteriori estimators; mixture models; parameter learning; prior balancing manipulation; trimming tests; unambiguous maximally structured models; Annealing; Bayesian methods; Entropy; Hidden Markov models; Maximum a posteriori estimation; Natural languages; Performance evaluation; Predictive models; Speech; Testing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech, and Signal Processing, 1999. Proceedings., 1999 IEEE International Conference on
Conference_Location :
Phoenix, AZ
ISSN :
1520-6149
Print_ISBN :
0-7803-5041-3
Type :
conf
DOI :
10.1109/ICASSP.1999.756333
Filename :
756333
Link To Document :
بازگشت