Title :
Fast linear discriminant analysis for on-line pattern recognition applications
Author :
Moghaddam, H. Abrishami ; Zadeh, Kh Amiri
Author_Institution :
K.N. Toossi Univ. of Technol., Tehran, Iran
Abstract :
In this paper, a new adaptive algorithm for Linear Discriminant Analysis (LDA) is presented. The major advantage of the algorithm is the fast convergence rate, which distinguishes it from the existing on-line methods. Current adaptive methods based on the gradient descent optimization technique use a fixed or a monotonically decreasing step size in each iteration. In this work, we use the steepest descent optimization method to optimally determine the step size in each iteration. It is shown that an optimally variable step size, significantly improves the convergence rate of the algorithm, compared to the conventional methods. The new algorithm has been implemented using a self-organized neural network and its advantages in on-line pattern recognition applications are demonstrated.
Keywords :
convergence of numerical methods; feature extraction; gradient methods; image recognition; self-organising feature maps; adaptive algorithm; fast convergence rate; fast linear discriminant analysis; iteration step size; on-line pattern recognition; optimally variable step size; self-organized neural network; steepest descent optimization method; Acceleration; Adaptive algorithm; Algorithm design and analysis; Convergence; Face detection; Feature extraction; Linear discriminant analysis; Optimization methods; Pattern recognition; Principal component analysis;
Conference_Titel :
Pattern Recognition, 2002. Proceedings. 16th International Conference on
Print_ISBN :
0-7695-1695-X
DOI :
10.1109/ICPR.2002.1048237