Title :
Regularization and error bars for the mixture of experts network
Author :
Ramamurti, Viswanath ; Ghosh, Joydeep
Author_Institution :
Dept. of Electr. & Comput. Eng., Texas Univ., Austin, TX, USA
Abstract :
The mixture of experts architecture provides a modular approach to function approximation. Since different experts get attuned to different regions of the input space during the course of training, and data distribution may not be uniform, some experts may get over-trained while others are undertrained. This leads to overall poorer generalization. In this paper, we show how regularization applied to the gating network improves generalization performance during the course of training. Secondly, we address the issue of estimating the error bars for network prediction. This is useful to estimate the range of probable network outputs for a given input especially in performance critical applications. Equations are derived to estimate the variance of the network output for a given input. Simulation results are presented in support of the proposed methods which substantially improve the effectiveness of mixture of experts networks
Keywords :
function approximation; neural net architecture; data distribution; error bars; experts mixture network; function approximation; modular approach; network prediction; performance critical applications; probable network outputs; regularization; Bars; Computer architecture; Computer errors; Contracts; Ear; Electronic mail; Equations; Function approximation; Lapping;
Conference_Titel :
Neural Networks,1997., International Conference on
Conference_Location :
Houston, TX
Print_ISBN :
0-7803-4122-8
DOI :
10.1109/ICNN.1997.611668