DocumentCode :
1944799
Title :
On practical use of stagewise second-order backpropagation for multi-stage neural-network learning
Author :
Mizutani, Eiji ; Dreyfus, Stuart
Author_Institution :
Nat. Taiwan Univ. of Sci. & Technol., Taipei
fYear :
2007
fDate :
12-17 Aug. 2007
Firstpage :
1302
Lastpage :
1307
Abstract :
We analyze the Hessian matrix H of the sum-squared-error measure for multilayer-perceptron (MLP) learning, showing the following intriguing results: At an early stage of learning, H is indefinite. The indefiniteness is related to the MLP structure, which also determines rank of H (per datum). Exploiting negative curvature leads to efficient learning algorithms. H can be much less ill-conditioned than the Gauss-Newton Hessian JTJ. All these new findings are obtained by our stagewise second-order backpropagation; the procedure exploits MLP\´s "layered symmetry" to evaluate H quickly, making exact Hessian evaluation feasible for fairly large practical problems. In fact, it works faster than rank-update methods that evaluate only JTJ. It also serves to appraise other existing learning algorithms that perform implicit Hessian-vector multiply.
Keywords :
backpropagation; multilayer perceptrons; Hessian-vector multiply; layered symmetry; multilayer-perceptron learning; multistage neural-network learning; rank-update methods; stagewise second-order backpropagation; sum-squared-error measure; Appraisal; Backpropagation algorithms; Cost function; Decision making; Gaussian processes; Jacobian matrices; Multi-layer neural network; Neural networks; Optimal control; Symmetric matrices;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2007. IJCNN 2007. International Joint Conference on
Conference_Location :
Orlando, FL
ISSN :
1098-7576
Print_ISBN :
978-1-4244-1379-9
Electronic_ISBN :
1098-7576
Type :
conf
DOI :
10.1109/IJCNN.2007.4371146
Filename :
4371146
Link To Document :
بازگشت