DocumentCode :
2644528
Title :
Improving the training speed of three-layer feedforward neural nets by optimal estimation of the initial weights
Author :
Chen, Chih-Liang ; Nutter, Roy S.
Author_Institution :
Dept. of Electr. & Comput. Eng., West Virginia Univ., Morgantown, WV, USA
fYear :
1991
fDate :
18-21 Nov 1991
Firstpage :
2063
Abstract :
The authors formulate the training problem for three-layer feedforward neural nets based on the well known linear algebra of D. Rumelhart et al. (1986). Then, they develop two estimation algorithms, called the forward estimation algorithm and the recurrent estimation algorithm, to estimate the initial weights. The basic idea is to set the initial weights space as close as possible to a global minimum before training, consequently reducing the training time. It is theoretically and empirically shown that a training procedure is unnecessary if the number of hidden units is equal to or greater than the number of training patterns minus one. Simulations were conducted for several problems. Results showed that the training speed is significantly improved by both initial weight estimation algorithms
Keywords :
learning systems; linear algebra; neural nets; optimisation; parallel algorithms; forward estimation algorithm; global minimum; learning systems; linear algebra; recurrent estimation algorithm; three-layer feedforward neural nets; weights space; Bismuth; Feedforward neural networks; Feedforward systems; Linear algebra; Neural networks; Random processes;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1991. 1991 IEEE International Joint Conference on
Print_ISBN :
0-7803-0227-3
Type :
conf
DOI :
10.1109/IJCNN.1991.170691
Filename :
170691
Link To Document :
بازگشت