DocumentCode
2957896
Title
The generalized back-propagation algorithm with convergence analysis
Author
Ng, S.C. ; Leung, S.H. ; Luk, A.
Author_Institution
Dept. of Comput. & Math., Hong Kong Tech. Coll., Hong Kong
Volume
5
fYear
1999
fDate
1999
Firstpage
612
Abstract
The conventional back-propagation algorithm is basically a gradient-descent method, it has the problems of local minima and slow convergence. The generalized back-propagation algorithm which can effectively speed up the convergence rate has been proposed previously. In this paper, the convergence proof of the algorithm is analyzed. The generalized backpropagation algorithm changes the derivative of the activation function so as to magnify the backward propagated error signal when the output approaches a wrong value; thus the convergence rate can be accelerated and the local minimum escaped. From the convergence analysis, it is shown that the generalized back-propagation algorithm will improve the original backpropagation algorithm in terms of faster convergence and global search capability
Keywords
backpropagation; convergence; generalisation (artificial intelligence); neural nets; activation function; backward propagated error signal; convergence analysis; generalized back-propagation algorithm; global search capability; neural nets; Acceleration; Algorithm design and analysis; Australia; Convergence; Educational institutions; Equations; Investments; Mathematics; Neural networks; Neurons;
fLanguage
English
Publisher
ieee
Conference_Titel
Circuits and Systems, 1999. ISCAS '99. Proceedings of the 1999 IEEE International Symposium on
Conference_Location
Orlando, FL
Print_ISBN
0-7803-5471-0
Type
conf
DOI
10.1109/ISCAS.1999.777646
Filename
777646
Link To Document