Title :
Shocking: an approach to stabilize backprop training with greedy adaptive learning rates
Author :
Janet, J.A. ; Scoggins, S.M. ; Schultz, S.M. ; Snyder, W.E. ; White, M.W. ; Sutton, J.C.
Author_Institution :
Dept. of Electr. & Comput. Eng., North Carolina State Univ., Raleigh, NC, USA
Abstract :
In general, backpropagation neural networks converge faster with adaptive learning rates than with learning rates that remain constant or grow or decay disregarding to the network error. This is because each synapse has its own learning rate that can vary over time by an amount appropriate to that weight. However, adaptive learning rates cause neural networks to saturate during training. When learning rates are permitted to assume values greater than unity, they are considered “greedy”. Greedy adaptive learning rates can reduce not only the training times of networks, but also compromise the stability of the training process, leading to a network that fails to converge. This paper proposes a simple ad hoc approach called “shocking” as a partial solution to the instability problem caused by greedy adaptive learning rates. An analysis based on training times and failure rates for two inherently unstable benchmark problems is used to validate the use of shocking
Keywords :
backpropagation; convergence; neural nets; stability; backpropagation; convergence; greedy adaptive learning rates; neural networks; stability; Computer errors; Electronic mail; Failure analysis; Jacobian matrices; Large-scale systems; Neural networks; Random variables; Stability; Time measurement;
Conference_Titel :
Neural Networks Proceedings, 1998. IEEE World Congress on Computational Intelligence. The 1998 IEEE International Joint Conference on
Conference_Location :
Anchorage, AK
Print_ISBN :
0-7803-4859-1
DOI :
10.1109/IJCNN.1998.687205