DocumentCode
2901430
Title
Training Single Hidden Layer Feedforward Neural Networks by Singular Value Decomposition
Author
Huynh, Hieu Trung ; Won, Yonggwan
Author_Institution
Dept. of Comput. Eng., Chonnam Nat. Univ., Gwangju, South Korea
fYear
2009
fDate
24-26 Nov. 2009
Firstpage
1300
Lastpage
1304
Abstract
Training neural networks has attracted many researchers for a long time. Many training algorithms and their improvements have been proposed. However, up to now, improving performance of training algorithms for neural networks is still a challenge. In this paper, we investigate a new training method for single hidden layer feedforward neural networks (SLFNs) which use `tansig´ activation function. The proposed training algorithm uses SVD (singular value decomposition) to calculate the network parameters. It is simple and has low computational complexity. Experimental results show that the proposed approach can obtain good performance with a compact network which has small number of hidden units.
Keywords
computational complexity; feedforward neural nets; learning (artificial intelligence); singular value decomposition; computational complexity; single hidden layer feedforward neural network training; singular value decomposition; tansig activation function; Computational complexity; Computer networks; Feedforward neural networks; Function approximation; Information technology; Least squares methods; Linear systems; Matrix decomposition; Neural networks; Singular value decomposition; SLFNs; SVD; SVD-neural classifier; neural network; training algorithms;
fLanguage
English
Publisher
ieee
Conference_Titel
Computer Sciences and Convergence Information Technology, 2009. ICCIT '09. Fourth International Conference on
Conference_Location
Seoul
Print_ISBN
978-1-4244-5244-6
Electronic_ISBN
978-0-7695-3896-9
Type
conf
DOI
10.1109/ICCIT.2009.170
Filename
5368523
Link To Document