Title :
Recursive Bayesian Levenberg-Marquardt Training of Recurrent Neural Networks
Author :
Mirikitani, Derrick ; Nikolaev, Nikolay
Author_Institution :
Univ. of London, London
Abstract :
This paper develops a Bayesian approach to recursive second order training of recurrent neural networks. A general recursive Levenberg-Marquardt algorithm is elaborated using Bayesian regularization. Individual local regularization hyperparameters as well as an output noise hyper-parameter are reestimated in order to maximize the weight posterior distribution and to produce a well generalizing network model. The proposed algorithm performs a computationally stable sequential Hessian estimation with RTRL derivatives. Experimental investigations using benchmark and practical data sets show that the developed algorithm outperforms the standard RTRL and extended Kalman training algorithms for recurrent nets, as well as feed forward and finite impulse response neural filters, on time series modeling.
Keywords :
FIR filters; Kalman filters; learning (artificial intelligence); recurrent neural nets; recursive estimation; sequential estimation; time series; RTRL algorithm; extended Kalman training algorithm; finite impulse response neural filter; local regularization hyperparameter; real time recurrent learning; recurrent neural networks; recursive Bayesian Levenberg-Marquardt training; stable sequential Hessian estimation; time series modeling; weight posterior distribution; Bayesian methods; Equations; Feeds; Filtering algorithms; Finite impulse response filter; Kalman filters; Neural networks; Recurrent neural networks; Standards development; Training data;
Conference_Titel :
Neural Networks, 2007. IJCNN 2007. International Joint Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
978-1-4244-1379-9
Electronic_ISBN :
1098-7576
DOI :
10.1109/IJCNN.2007.4370969