DocumentCode :
1909397
Title :
Generalization and maximum likelihood from small data sets
Author :
Byrne, William
Author_Institution :
Dept. of Electr. Eng., Maryland Univ., College Park, MD, USA
fYear :
1993
fDate :
6-9 Sep 1993
Firstpage :
197
Lastpage :
206
Abstract :
A technique is described which can be used to prevent overtraining and encourage generalization in training under a maximum likelihood criterion. Applications to Boltzmann machines and hidden Markov models (HMMs) are discussed. While the confidence constraint may slow the training algorithm, in general it should involve very little additional calculation. The results presented for HMMs are for training under a maximum likelihood criterion based on the marginal distribution. Similar modifications can be made to the segmental K-means and N-best algorithms
Keywords :
Boltzmann machines; generalisation (artificial intelligence); hidden Markov models; learning (artificial intelligence); neural nets; Boltzmann machines; K-means algorithm; N-best algorithms; generalization; hidden Markov models; maximum likelihood criterion; Counting circuits; Distributed computing; Educational institutions; Hidden Markov models; Iterative algorithms; Probability; Q measurement; Quadratic programming; Statistical distributions; Training data;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks for Processing [1993] III. Proceedings of the 1993 IEEE-SP Workshop
Conference_Location :
Linthicum Heights, MD
Print_ISBN :
0-7803-0928-6
Type :
conf
DOI :
10.1109/NNSP.1993.471869
Filename :
471869
Link To Document :
بازگشت