• DocumentCode
    768096
  • Title

    Dynamic learning rate optimization of the backpropagation algorithm

  • Author

    Yu, Xiao-Hu ; Chen, Guo-An ; Cheng, Shi-xin

  • Author_Institution
    Dept. of Radio Eng., Southeast Univ., Nanjing, China
  • Volume
    6
  • Issue
    3
  • fYear
    1995
  • fDate
    5/1/1995 12:00:00 AM
  • Firstpage
    669
  • Lastpage
    677
  • Abstract
    It has been observed by many authors that the backpropagation (BP) error surfaces usually consist of a large amount of flat regions as well as extremely steep regions. As such, the BP algorithm with a fixed learning rate will have low efficiency. This paper considers dynamic learning rate optimization of the BP algorithm using derivative information. An efficient method of deriving the first and second derivatives of the objective function with respect to the learning rate is explored, which does not involve explicit calculation of second-order derivatives in weight space, but rather uses the information gathered from the forward and backward propagation, Several learning rate optimization approaches are subsequently established based on linear expansion of the actual outputs and line searches with acceptable descent value and Newton-like methods, respectively. Simultaneous determination of the optimal learning rate and momentum is also introduced by showing the equivalence between the momentum version BP and the conjugate gradient method. Since these approaches are constructed by simple manipulations of the obtained derivatives, the computational and storage burden scale with the network size exactly like the standard BP algorithm, and the convergence of the BP algorithm is accelerated with in a remarkable reduction (typically by factor 10 to 50, depending upon network architectures and applications) in the running time for the overall learning process. Numerous computer simulation results are provided to support the present approaches
  • Keywords
    Newton method; backpropagation; conjugate gradient methods; convergence; neural nets; optimisation; Newton-like method; backpropagation algorithm; backpropagation error surfaces; backward propagation; conjugate gradient method; convergence; derivative information; descent value method; dynamic learning rate optimization; flat regions; forward propagation; line searches; linear expansion; momentum; steep regions; Acceleration; Application software; Backpropagation algorithms; Computer architecture; Computer networks; Computer simulation; Convergence; Gradient methods; Iterative algorithms; Optimization methods;
  • fLanguage
    English
  • Journal_Title
    Neural Networks, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1045-9227
  • Type

    jour

  • DOI
    10.1109/72.377972
  • Filename
    377972