DocumentCode
311201
Title
A neural network training algorithm utilizing multiple sets of linear equations
Author
Chen, Hung-Han ; Manry, Michael T.
Author_Institution
Dept. of Electr. Eng., Texas Univ., Arlington, TX, USA
fYear
1996
fDate
3-6 Nov. 1996
Firstpage
1166
Abstract
A fast algorithm is presented for the training of multilayer perceptron neural networks. In each iteration, there are two passes through the training data. In the first pass, linear equations are solved for the output weights. In the second data pass, linear equations are solved for hidden unit weight changes. Full batching is used in both data passes. An algorithm is described for calculating the learning factor for use with the hidden weights. It is shown that the technique is significantly faster than standard output weight optimization-backpropagation.
Keywords
learning (artificial intelligence); multilayer perceptrons; batching; data passes; fast algorithm; hidden unit weight; learning factor; linear equations; multilayer perceptron neural networks; neural network training algorithm; output weight optimization-backpropagation; output weights; surface scattering parameters; training data; Additives; Backpropagation algorithms; Equations; Feeds; Joining processes; Mean square error methods; Multi-layer neural network; Multilayer perceptrons; Neural networks; Training data;
fLanguage
English
Publisher
ieee
Conference_Titel
Signals, Systems and Computers, 1996. Conference Record of the Thirtieth Asilomar Conference on
Conference_Location
Pacific Grove, CA, USA
ISSN
1058-6393
Print_ISBN
0-8186-7646-9
Type
conf
DOI
10.1109/ACSSC.1996.599128
Filename
599128
Link To Document