DocumentCode :
313582
Title :
Regularization and error bars for the mixture of experts network
Author :
Ramamurti, Viswanath ; Ghosh, Joydeep
Author_Institution :
Dept. of Electr. & Comput. Eng., Texas Univ., Austin, TX, USA
Volume :
1
fYear :
1997
fDate :
9-12 Jun 1997
Firstpage :
221
Abstract :
The mixture of experts architecture provides a modular approach to function approximation. Since different experts get attuned to different regions of the input space during the course of training, and data distribution may not be uniform, some experts may get over-trained while others are undertrained. This leads to overall poorer generalization. In this paper, we show how regularization applied to the gating network improves generalization performance during the course of training. Secondly, we address the issue of estimating the error bars for network prediction. This is useful to estimate the range of probable network outputs for a given input especially in performance critical applications. Equations are derived to estimate the variance of the network output for a given input. Simulation results are presented in support of the proposed methods which substantially improve the effectiveness of mixture of experts networks
Keywords :
function approximation; neural net architecture; data distribution; error bars; experts mixture network; function approximation; modular approach; network prediction; performance critical applications; probable network outputs; regularization; Bars; Computer architecture; Computer errors; Contracts; Ear; Electronic mail; Equations; Function approximation; Lapping;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks,1997., International Conference on
Conference_Location :
Houston, TX
Print_ISBN :
0-7803-4122-8
Type :
conf
DOI :
10.1109/ICNN.1997.611668
Filename :
611668
Link To Document :
بازگشت