• DocumentCode
    2957896
  • Title

    The generalized back-propagation algorithm with convergence analysis

  • Author

    Ng, S.C. ; Leung, S.H. ; Luk, A.

  • Author_Institution
    Dept. of Comput. & Math., Hong Kong Tech. Coll., Hong Kong
  • Volume
    5
  • fYear
    1999
  • fDate
    1999
  • Firstpage
    612
  • Abstract
    The conventional back-propagation algorithm is basically a gradient-descent method, it has the problems of local minima and slow convergence. The generalized back-propagation algorithm which can effectively speed up the convergence rate has been proposed previously. In this paper, the convergence proof of the algorithm is analyzed. The generalized backpropagation algorithm changes the derivative of the activation function so as to magnify the backward propagated error signal when the output approaches a wrong value; thus the convergence rate can be accelerated and the local minimum escaped. From the convergence analysis, it is shown that the generalized back-propagation algorithm will improve the original backpropagation algorithm in terms of faster convergence and global search capability
  • Keywords
    backpropagation; convergence; generalisation (artificial intelligence); neural nets; activation function; backward propagated error signal; convergence analysis; generalized back-propagation algorithm; global search capability; neural nets; Acceleration; Algorithm design and analysis; Australia; Convergence; Educational institutions; Equations; Investments; Mathematics; Neural networks; Neurons;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Circuits and Systems, 1999. ISCAS '99. Proceedings of the 1999 IEEE International Symposium on
  • Conference_Location
    Orlando, FL
  • Print_ISBN
    0-7803-5471-0
  • Type

    conf

  • DOI
    10.1109/ISCAS.1999.777646
  • Filename
    777646