Title :
Averaging, maximum penalized likelihood and Bayesian estimation for improving Gaussian mixture probability density estimates
Author :
Ormoneit, Dirk ; Tresp, Volker
Author_Institution :
Dept. of Comput. Sci., Tech. Univ. Munchen, Germany
fDate :
7/1/1998 12:00:00 AM
Abstract :
We apply the idea of averaging ensembles of estimators to probability density estimation. In particular, we use Gaussian mixture models which are important components in many neural-network applications. We investigate the performance of averaging using three data sets. For comparison, we employ two traditional regularization approaches, i.e., a maximum penalized likelihood approach and a Bayesian approach. In the maximum penalized likelihood approach we use penalty functions derived from conjugate Bayesian priors such that an expectation maximization (EM) algorithm can be used for training. In all experiments, the maximum penalized likelihood approach and averaging improved performance considerably if compared to a maximum likelihood approach. In two of the experiments, the maximum penalized likelihood approach outperformed averaging. In one experiment averaging was clearly superior. Our conclusion is that maximum penalized likelihood gives good results if the penalty term in the cost function is appropriate for the particular problem. If this is not the case, averaging is superior since it shows greater robustness by not relying on any particular prior assumption. The Bayesian approach worked very well on a low-dimensional toy problem but failed to give good performance in higher dimensional problems
Keywords :
Bayes methods; Gaussian distribution; maximum likelihood estimation; neural nets; probability; Bayesian estimation; EM algorithm; Gaussian mixture probability density estimates; averaging; conjugate Bayesian priors; expectation maximization algorithm; maximum penalized likelihood; penalty functions; penalty term; regularization; robustness; Bagging; Bayesian methods; Cost function; Inference algorithms; Maximum likelihood estimation; Neural networks; Robustness; Sampling methods; Space technology; Training data;
Journal_Title :
Neural Networks, IEEE Transactions on