DocumentCode
3009563
Title
Robust neural network training using partial gradient probing
Author
Manic, Milos ; Wilamowski, Bogdan
Author_Institution
Dept. of Comput. Sci., Idaho Univ., Boise, ID, USA
fYear
2003
fDate
21-24 Aug. 2003
Firstpage
175
Lastpage
180
Abstract
Our proposed algorithm features fast and robust convergence for one hidden layer neural networks. Search for weights is done only in the input layer i.e. on compressed network. Only forward propagation is performed with second layer trained automatically with pseudo-inversion training, for all patterns at once. Last layer training is also modified to handle nonlinear problems, not presented here. Through iterations gradient is randomly probed towards each weight set dimension. The algorithm further features serious of modifications, such as adaptive network parameters that resolve problems like total error fluctuations, slow convergence, etc. For testing of this algorithm one of most popular benchmark tests - parity problems were chosen. Final version of the proposed algorithm typically provides a solution for various tested parity problems in less than ten iterations, regardless of initial weight set. Performance of the algorithm on parity problems is tested and illustrated by figures.
Keywords
gradient methods; learning (artificial intelligence); multilayer perceptrons; neural net architecture; search problems; forward propagation; hidden layer neural network; neural net architecture; neural network training; parity problem; partial gradient probing; pseudo-inversion training; Computer science; Convergence; Fluctuations; Iterative algorithms; Iterative methods; Neural networks; Neurons; Robustness; Search methods; Testing;
fLanguage
English
Publisher
ieee
Conference_Titel
Industrial Informatics, 2003. INDIN 2003. Proceedings. IEEE International Conference on
Print_ISBN
0-7803-8200-5
Type
conf
DOI
10.1109/INDIN.2003.1300266
Filename
1300266
Link To Document