Title :
A Normalized Adaptive Training of Recurrent Neural Networks With Augmented Error Gradient
Author :
Yilei, Wu ; Qing, Song ; Sheng, Liu
Author_Institution :
Nanyang Technol. Univ., Singapore
Abstract :
For training algorithms of recurrent neural networks (RNN), convergent speed and training error are always two contradictory performances. In this letter, we propose a normalized adaptive recurrent learning (NARL) to obtain a tradeoff between transient and steady-state response. An augmented term is added to error gradient to exactly model the derivative of cost function with respect to hidden layer weight. The influence of the induced gain of activation function on training stability is also taken into consideration. Moreover, adaptive learning rate is employed to improve the robustness of the gradient training. Finally, computer simulations of a model prediction problem are synthesized to give comparisons between NARL and conventional normalized real-time recurrent learning (N-RTRL).
Keywords :
adaptive systems; error analysis; gradient methods; learning (artificial intelligence); recurrent neural nets; stability; augmented error gradient; normalized adaptive recurrent learning; recurrent neural networks; steady-state response; training error; training stability; Adaptive learning rate; augmented error gradient; convergence; normalization; Adaptation, Biological; Algorithms; Computer Simulation; Humans; Learning; Neural Networks (Computer); Signal Processing, Computer-Assisted; Time Factors;
Journal_Title :
Neural Networks, IEEE Transactions on
DOI :
10.1109/TNN.2007.908647