Title :
Sparse support vector machines trained in the reduced empirical feature space
Author :
Iwamura, Kazuki ; Abe, Shigeo
Author_Institution :
Electr. Eng., Kobe Univ., Kobe
Abstract :
We discuss sparse support vector machines (sparse SVMs) trained in the reduced empirical feature space. Namely, we select the linearly independent training data by the Cholesky factorization of the kernel matrix, and train the SVM in the dual form in the reduced empirical feature space. Since the mapped linearly independent training data span the empirical feature space, the linearly independent training data become support vectors. Thus if the number of linearly independent data is smaller than the number of support vectors trained in the feature space, sparsity is increased. By computer experiments we show that in most cases we can reduce the number of support vectors without deteriorating the generalization ability.
Keywords :
matrix decomposition; support vector machines; Cholesky factorization; independent training data; reduced empirical feature space; sparse support vector machines; Constraint optimization; Eigenvalues and eigenfunctions; Kernel; Least squares methods; Newton method; Sparse matrices; Support vector machine classification; Support vector machines; Symmetric matrices; Training data;
Conference_Titel :
Neural Networks, 2008. IJCNN 2008. (IEEE World Congress on Computational Intelligence). IEEE International Joint Conference on
Conference_Location :
Hong Kong
Print_ISBN :
978-1-4244-1820-6
Electronic_ISBN :
1098-7576
DOI :
10.1109/IJCNN.2008.4634131