Title :
Nonlinear kernel-based statistical pattern analysis
Author :
Ruiz, Alberto ; López-de-Teruel, Pedro E.
Author_Institution :
Dept. of Comput. Sci., Murcia Univ., Spain
fDate :
1/1/2001 12:00:00 AM
Abstract :
The eigenstructure of the second-order statistics of a multivariate random population can be inferred from the matrix of pairwise combinations of inner products of the samples. Therefore, it can be also efficiently obtained in the implicit, high-dimensional feature spaces defined by kernel functions. We elaborate on this property to obtain general expressions for immediate derivation of nonlinear counterparts of a number of standard pattern analysis algorithms, including principal component analysis, data compression and denoising, and Fisher´s discriminant. The connection between kernel methods and nonparametric density estimation is also illustrated. Using these results we introduce the kernel version of Mahalanobis distance, which originates nonparametric models with unexpected and interesting properties, and also propose a kernel version of the minimum squared error (MSE) linear discriminant function. This learning machine is particularly simple and includes a number of generalized linear models such as the potential functions method or the radial basis function (RBF) network. Our results shed some light on the relative merit of feature spaces and inductive bias in the remarkable generalization properties of the support vector machine (SVM). Although in most situations the SVM obtains the lowest error rates, exhaustive experiments with synthetic and natural data show that simple kernel machines based on pseudoinversion are competitive in problems with appreciable class overlapping
Keywords :
covariance matrices; data compression; eigenvalues and eigenfunctions; feature extraction; learning automata; learning systems; mean square error methods; nonparametric statistics; pattern classification; principal component analysis; Fisher´s discriminant; Mahalanobis distance; data compression; denoising; eigenstructure; feature spaces; generalization properties; generalized linear models; implicit high-dimensional feature spaces; inductive bias; learning machine; minimum squared error linear discriminant function; multivariate random population; nonlinear kernel-based statistical pattern analysis; nonparametric density estimation; pairwise combinations; potential functions method; principal component analysis; second-order statistics; support vector machine; Data compression; Error analysis; Genetic expression; Kernel; Machine learning; Noise reduction; Pattern analysis; Principal component analysis; Statistics; Support vector machines;
Journal_Title :
Neural Networks, IEEE Transactions on