DocumentCode
445932
Title
Evaluation of cluster combination functions for mixture of experts
Author
Redhead, Robert ; Heywood, Malcolm
Author_Institution
Fac. of Comput. Sci., Dalhousie Univ., Halifax, NS, Canada
Volume
2
fYear
2005
fDate
31 July-4 Aug. 2005
Firstpage
1154
Abstract
The mixtures of experts (MoE) model provides the basis for building modular neural network solutions. In this work we are interested in methods for decomposing the input before forwarding to the MoE architecture. By doing so we are able to define the number of experts from the data itself. Specific schemes are shown to be appropriate for regression and classification problems, where each appear to have different preferences.
Keywords
neural nets; pattern classification; classification problems; cluster combination functions; mixtures of experts model; modular neural network; regression problems; Buildings; Clustering algorithms; Computer architecture; Computer science; Data preprocessing; Jacobian matrices; Neural networks; Neurons; Partitioning algorithms; Switches;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 2005. IJCNN '05. Proceedings. 2005 IEEE International Joint Conference on
Print_ISBN
0-7803-9048-2
Type
conf
DOI
10.1109/IJCNN.2005.1556016
Filename
1556016
Link To Document