DocumentCode :
2109758
Title :
Learning laws with exponential error convergence for recurrent neural networks
Author :
Kosmatopoulos, Elias B. ; Christdoulou, M.A. ; Ioannou, Peters A.
Author_Institution :
Dept. of Electron. & Comput. Eng., Tech. Univ. of Crete, Chania, Greece
fYear :
1993
fDate :
15-17 Dec 1993
Firstpage :
2810
Abstract :
In this paper, we propose new learning laws for adjusting the weights of recurrent high order neural networks (RHONN) when they are used to system identification problems. The main advantages of these learning laws over the classical robust adaptive ones, is that the identification error converges to zero exponentially fast, and that such a convergence is independent of the number of high order connections of the RHONN
Keywords :
convergence of numerical methods; error analysis; identification; learning (artificial intelligence); nonlinear systems; recurrent neural nets; exponential error convergence; learning laws; nonlinear systems; recurrent neural networks; system identification; Adaptive algorithm; Adaptive control; Convergence; Lyapunov method; Neural networks; Neurons; Parameter estimation; Programmable control; Recurrent neural networks; Robustness;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Decision and Control, 1993., Proceedings of the 32nd IEEE Conference on
Conference_Location :
San Antonio, TX
Print_ISBN :
0-7803-1298-8
Type :
conf
DOI :
10.1109/CDC.1993.325707
Filename :
325707
Link To Document :
بازگشت