DocumentCode
464891
Title
Generalizations of Oja´s Learning Rule to Non-Symmetric Matrices
Author
Hasan, Mohammed A.
Author_Institution
Dept. of Electr. & Comput. Eng., Minnesota Univ., Duluth, MN
fYear
2007
fDate
27-30 May 2007
Firstpage
1779
Lastpage
1782
Abstract
New learning rules for computing eigenspaces and eigenvectors for symmetric and nonsymmetric matrices are proposed. By applying Liapunov stability theory, these systems are shown to be globally convergent. Properties of limiting solutions of the systems and weighted versions are also examined. The proposed systems may be viewed as generalizations of Oja´s and Xu´s principal subspace learning rules. Numerical examples showing the convergence behavior are also presented.
Keywords
eigenvalues and eigenfunctions; learning (artificial intelligence); numerical stability; principal component analysis; Liapunov stability theory; Oja learning rule; convergence behavior; eigenspaces computing; eigenvectors computing; global convergence; limiting solutions; minor components; nonsymmetric matrices; principal components; principal subspace learning rules; symmetric matrices; weighted versions; Convergence of numerical methods; Eigenvalues and eigenfunctions; Lagrangian functions; Lyapunov method; Principal component analysis; Signal analysis; Signal processing; Signal processing algorithms; Stability; Symmetric matrices; Liapunov stability; Oja´s learning rule; global convergence; minor components; principal components;
fLanguage
English
Publisher
ieee
Conference_Titel
Circuits and Systems, 2007. ISCAS 2007. IEEE International Symposium on
Conference_Location
New Orleans, LA
Print_ISBN
1-4244-0920-9
Electronic_ISBN
1-4244-0921-7
Type
conf
DOI
10.1109/ISCAS.2007.378017
Filename
4253004
Link To Document