DocumentCode :
1509900
Title :
An analysis of noise in recurrent neural networks: convergence and generalization
Author :
Jim, Kam-Chuen ; Giles, C. Lee ; Horne, Bill G.
Author_Institution :
NEC Res. Inst., Princeton, NJ, USA
Volume :
7
Issue :
6
fYear :
1996
fDate :
11/1/1996 12:00:00 AM
Firstpage :
1424
Lastpage :
1438
Abstract :
Concerns the effect of noise on the performance of feedforward neural nets. We introduce and analyze various methods of injecting synaptic noise into dynamically driven recurrent nets during training. Theoretical results show that applying a controlled amount of noise during training may improve convergence and generalization performance. We analyze the effects of various noise parameters and predict that best overall performance can be achieved by injecting additive noise at each time step. Noise contributes a second-order gradient term to the error function which can be viewed as an anticipatory agent to aid convergence. This term appears to find promising regions of weight space in the beginning stages of training when the training error is large and should improve convergence on error surfaces with local minima. The first-order term is a regularization term that can improve generalization. Specifically, it can encourage internal representations where the state nodes operate in the saturated regions of the sigmoid discriminant function. While this effect can improve performance on automata inference problems with binary inputs and target outputs, it is unclear what effect it will have on other types of problems. To substantiate these predictions, we present simulations on learning the dual parity grammar from temporal strings for all noise models, and present simulations on learning a randomly generated six-state grammar using the predicted best noise model
Keywords :
automata theory; convergence; duality (mathematics); feedforward neural nets; generalisation (artificial intelligence); grammars; inference mechanisms; noise; recurrent neural nets; automata inference problems; convergence; dual parity grammar learning; dynamically driven recurrent nets; feedforward neural nets; generalization; noise analysis; randomly generated six-state grammar; recurrent neural networks; saturated regions; second-order gradient term; sigmoid discriminant function; synaptic noise; temporal strings; training; weight space; Additive noise; Central Processing Unit; Convergence; Feedforward neural networks; Intelligent networks; Neural networks; Noise generators; Predictive models; State-space methods; Temperature;
fLanguage :
English
Journal_Title :
Neural Networks, IEEE Transactions on
Publisher :
ieee
ISSN :
1045-9227
Type :
jour
DOI :
10.1109/72.548170
Filename :
548170
Link To Document :
بازگشت