Title :
Convex error functions for multilayered perceptrons
Author_Institution :
Ecole Normale Superieure, Paris, France
Abstract :
For multilayered perceptrons, the error is generally defined as a quadratic function of the difference between the ideal and real outputs. This L2 error, considered as a function of the weights, is generally not convex. The author studies the convexity properties of more general error functions for perceptrons with 0 or more hidden layers. A necessary and sufficient condition is presented for the convexity of the error function for all perceptrons without a hidden layer, and an explicit solution is given. A negative result for the convexity for perceptrons is included with one or more than one hidden layers. It is shown how the loss of global convexity for general perceptrons can advantageously be solved in practical applications by the use of a fitted minimization algorithm. An application to the problem of recognition of printed characters is included
Keywords :
error analysis; minimisation; neural nets; L2 error; convex error functions; fitted minimization algorithm; hidden layers; multilayered perceptrons; necessary and sufficient condition; perceptron convexity; printed character recognition; quadratic function; weights; Backpropagation algorithms; Character recognition; Convergence; Multilayer perceptrons;
Conference_Titel :
Neural Networks, 1991. 1991 IEEE International Joint Conference on
Print_ISBN :
0-7803-0227-3
DOI :
10.1109/IJCNN.1991.170770