DocumentCode
1301534
Title
Upper bounds on the number of hidden neurons in feedforward networks with arbitrary bounded nonlinear activation functions
Author
Huang, Guang-Bin ; Babri, Haroon A.
Author_Institution
Sch. of Electr. & Electron. Eng., Nanyang Technol. Univ., Singapore
Volume
9
Issue
1
fYear
1998
fDate
1/1/1998 12:00:00 AM
Firstpage
224
Lastpage
229
Abstract
It is well known that standard single-hidden layer feedforward networks (SLFNs) with at most N hidden neurons (including biases) can learn N distinct samples (xi,ti) with zero error, and the weights connecting the input neurons and the hidden neurons can be chosen “almost” arbitrarily. However, these results have been obtained for the case when the activation function for the hidden neurons is the signum function. This paper rigorously proves that standard single-hidden layer feedforward networks (SLFNs) with at most N hidden neurons and with any bounded nonlinear activation function which has a limit at one infinity can learn N distinct samples (xi,ti) with zero error. The previous method of arbitrarily choosing weights is not feasible for any SLFN. The proof of our result is constructive and thus gives a method to directly find the weights of the standard SLFNs with any such bounded nonlinear activation function as opposed to iterative training algorithms in the literature
Keywords
feedforward neural nets; learning (artificial intelligence); transfer functions; arbitrary bounded nonlinear activation functions; hidden neurons; single-hidden layer feedforward networks; upper bounds; Artificial neural networks; Feedforward neural networks; H infinity control; Intelligent networks; Iterative algorithms; Iterative methods; Joining processes; Multi-layer neural network; Neural networks; Neurons;
fLanguage
English
Journal_Title
Neural Networks, IEEE Transactions on
Publisher
ieee
ISSN
1045-9227
Type
jour
DOI
10.1109/72.655045
Filename
655045
Link To Document