Title :
Local gain adaptation in stochastic gradient descent
Author :
Schraudolph, Nicol N.
Author_Institution :
IDSIA, Lugano, Switzerland
Abstract :
Gain adaptation algorithms for neural networks typically adjust learning rates by monitoring the correlation between successive gradients. Here we discuss the limitations of this approach, and develop an alternative by extending Sutton´s work on linear systems (1992) to the general, nonlinear case. The resulting online algorithms are computationally little more expensive than other acceleration techniques, do not assume statistical independence between successive training patterns, and do not require an arbitrary smoothing parameter. In our benchmark experiments, they consistently outperform other acceleration methods, and show remarkable robustness when faced with non i.i.d. sampling of the input space
Keywords :
gradient methods; gradient correlation monitoring; learning rate adjustment; local gain adaptation; neural networks; non i.i.d. sampling; nonlinear systems; online algorithms; stochastic gradient descent;
Conference_Titel :
Artificial Neural Networks, 1999. ICANN 99. Ninth International Conference on (Conf. Publ. No. 470)
Conference_Location :
Edinburgh
Print_ISBN :
0-85296-721-7
DOI :
10.1049/cp:19991170