DocumentCode :
1340564
Title :
Comments on "Accelerated learning algorithm for multilayer perceptrons: optimization layer by layer"
Author :
Van Milligen, B.Ph. ; Tribaldos, V. ; Jimenez, J.A. ; Santa Cruz, C.
Author_Institution :
Inst. de Ingenieria des Conocimiento, Asociacion EURATOM, Madrid, Spain
Volume :
9
Issue :
2
fYear :
1998
fDate :
3/1/1998 12:00:00 AM
Firstpage :
339
Lastpage :
341
Abstract :
In the above paper by Ergezinger and Thomsen (ibid. vol.6 (1991)), a new method for training multilayer perceptron, called optimization layer by layer (OLL), was introduced. The present paper analyzes the performance of OLL. We show, from theoretical considerations, that the amount of work required with OLL-learning scales as the third power of the network size, compared with the square of the network size for commonly used conjugate gradient (CG) training algorithms. This theoretical estimate is confirmed through a practical example. Thus, although OLL is shown to function very well for small neural networks (less than about 500 weights per layer), it is slower than CG for large neural networks. Next, we show that OLL does not always improve on the accuracy that can be obtained with CG. It seems that the final accuracy that can be obtained depends strongly on the initial network weights.
Keywords :
learning (artificial intelligence); multilayer perceptrons; optimisation; performance evaluation; accelerated learning algorithm; conjugate gradient; multilayer perceptrons; optimization layer by layer; performance evaluation; Acceleration; Algorithm design and analysis; Character generation; Equations; Matrix decomposition; Multi-layer neural network; Multilayer perceptrons; Neural networks; Optimization methods; Performance analysis;
fLanguage :
English
Journal_Title :
Neural Networks, IEEE Transactions on
Publisher :
ieee
ISSN :
1045-9227
Type :
jour
DOI :
10.1109/72.661128
Filename :
661128
Link To Document :
بازگشت