Title :
Kernel Uncorrelated and Regularized Discriminant Analysis: A Theoretical and Computational Study
Author :
Ji, Shuiwang ; Ye, Jieping
Author_Institution :
Arizona State Univ., Tempe, AZ
Abstract :
Linear and kernel discriminant analyses are popular approaches for supervised dimensionality reduction. Uncorrelated and regularized discriminant analyses have been proposed to overcome the singularity problem encountered by classical discriminant analysis. In this paper, we study the properties of kernel uncorrelated and regularized discriminant analyses, called KUDA and KRDA, respectively. In particular, we show that under a mild condition, both linear and kernel uncorrelated discriminant analysis project samples in the same class to a common vector in the dimensionality-reduced space. This implies that uncorrelated discriminant analysis may suffer from the overfitting problem if there are a large number of samples in each class. We show that as the regularization parameter in KRDA tends to zero, KRDA approaches KUDA. This shows that KUDA is a special case of KRDA and that regularization can be applied to overcome the overfitting problem in uncorrelated discriminant analysis. As the performance of KRDA depends on the value of the regularization parameter, we show that the matrix computations involved in KRDA can be simplified, so that a large number of candidate values can be cross-validated efficiently. Finally, we conduct experiments to evaluate the proposed theories and algorithms.
Keywords :
data handling; matrix algebra; KRDA; KUDA; kernel uncorrelated discriminant analysis; linear discriminant analyses; matrix computations; overfitting problem; regularized discriminant analysis; singularity problem; supervised dimensionality reduction; Eigenvalues and eigenvectors; Feature extraction or construction; Parameter learning; Singular value decomposition;
Journal_Title :
Knowledge and Data Engineering, IEEE Transactions on
DOI :
10.1109/TKDE.2008.57