Title :
On practical use of stagewise second-order backpropagation for multi-stage neural-network learning
Author :
Mizutani, Eiji ; Dreyfus, Stuart
Author_Institution :
Nat. Taiwan Univ. of Sci. & Technol., Taipei
Abstract :
We analyze the Hessian matrix H of the sum-squared-error measure for multilayer-perceptron (MLP) learning, showing the following intriguing results: At an early stage of learning, H is indefinite. The indefiniteness is related to the MLP structure, which also determines rank of H (per datum). Exploiting negative curvature leads to efficient learning algorithms. H can be much less ill-conditioned than the Gauss-Newton Hessian JTJ. All these new findings are obtained by our stagewise second-order backpropagation; the procedure exploits MLP\´s "layered symmetry" to evaluate H quickly, making exact Hessian evaluation feasible for fairly large practical problems. In fact, it works faster than rank-update methods that evaluate only JTJ. It also serves to appraise other existing learning algorithms that perform implicit Hessian-vector multiply.
Keywords :
backpropagation; multilayer perceptrons; Hessian-vector multiply; layered symmetry; multilayer-perceptron learning; multistage neural-network learning; rank-update methods; stagewise second-order backpropagation; sum-squared-error measure; Appraisal; Backpropagation algorithms; Cost function; Decision making; Gaussian processes; Jacobian matrices; Multi-layer neural network; Neural networks; Optimal control; Symmetric matrices;
Conference_Titel :
Neural Networks, 2007. IJCNN 2007. International Joint Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
978-1-4244-1379-9
Electronic_ISBN :
1098-7576
DOI :
10.1109/IJCNN.2007.4371146