Title :
Optimal learning rates for each pattern and neuron in gradient descent training of multilayer perceptrons
Author :
Oh, Sang-Hoon ; Lee, Soo-Young
Author_Institution :
Dept. of Electr. Eng., Korea Adv. Inst. of Sci. & Technol., Taejon, South Korea
Abstract :
Proposes optimal learning rates in gradient descent training of multilayer perceptrons, which are a separate learning rate for weights associated with each neuron and a separate one for assigning virtual hidden targets associated with each training pattern. That is, a hidden weight vector has two optimal learning rates, one for assigning virtual hidden targets and the other for minimizing a hidden error function proposed in the paper. Effectiveness of the proposed error function was demonstrated for handwritten digit recognition and an isolated-word recognition tasks and very fast learning convergence was obtained
Keywords :
convergence; gradient methods; handwritten character recognition; learning (artificial intelligence); multilayer perceptrons; speech recognition; gradient descent training; handwritten digit recognition; hidden error function; isolated-word recognition; optimal learning rates; training pattern; very fast learning convergence; virtual hidden targets; Acceleration; Backpropagation algorithms; Computational complexity; Convergence; Handwriting recognition; Matrix decomposition; Multilayer perceptrons; Neurons; Nonhomogeneous media; Optimization methods;
Conference_Titel :
Neural Networks, 1999. IJCNN '99. International Joint Conference on
Conference_Location :
Washington, DC
Print_ISBN :
0-7803-5529-6
DOI :
10.1109/IJCNN.1999.832617