• DocumentCode
    1944799
  • Title

    On practical use of stagewise second-order backpropagation for multi-stage neural-network learning

  • Author

    Mizutani, Eiji ; Dreyfus, Stuart

  • Author_Institution
    Nat. Taiwan Univ. of Sci. & Technol., Taipei
  • fYear
    2007
  • fDate
    12-17 Aug. 2007
  • Firstpage
    1302
  • Lastpage
    1307
  • Abstract
    We analyze the Hessian matrix H of the sum-squared-error measure for multilayer-perceptron (MLP) learning, showing the following intriguing results: At an early stage of learning, H is indefinite. The indefiniteness is related to the MLP structure, which also determines rank of H (per datum). Exploiting negative curvature leads to efficient learning algorithms. H can be much less ill-conditioned than the Gauss-Newton Hessian JTJ. All these new findings are obtained by our stagewise second-order backpropagation; the procedure exploits MLP\´s "layered symmetry" to evaluate H quickly, making exact Hessian evaluation feasible for fairly large practical problems. In fact, it works faster than rank-update methods that evaluate only JTJ. It also serves to appraise other existing learning algorithms that perform implicit Hessian-vector multiply.
  • Keywords
    backpropagation; multilayer perceptrons; Hessian-vector multiply; layered symmetry; multilayer-perceptron learning; multistage neural-network learning; rank-update methods; stagewise second-order backpropagation; sum-squared-error measure; Appraisal; Backpropagation algorithms; Cost function; Decision making; Gaussian processes; Jacobian matrices; Multi-layer neural network; Neural networks; Optimal control; Symmetric matrices;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 2007. IJCNN 2007. International Joint Conference on
  • Conference_Location
    Orlando, FL
  • ISSN
    1098-7576
  • Print_ISBN
    978-1-4244-1379-9
  • Electronic_ISBN
    1098-7576
  • Type

    conf

  • DOI
    10.1109/IJCNN.2007.4371146
  • Filename
    4371146