DocumentCode :
2959862
Title :
Sparse support vector machines trained in the reduced empirical feature space
Author :
Iwamura, Kazuki ; Abe, Shigeo
Author_Institution :
Electr. Eng., Kobe Univ., Kobe
fYear :
2008
fDate :
1-8 June 2008
Firstpage :
2398
Lastpage :
2404
Abstract :
We discuss sparse support vector machines (sparse SVMs) trained in the reduced empirical feature space. Namely, we select the linearly independent training data by the Cholesky factorization of the kernel matrix, and train the SVM in the dual form in the reduced empirical feature space. Since the mapped linearly independent training data span the empirical feature space, the linearly independent training data become support vectors. Thus if the number of linearly independent data is smaller than the number of support vectors trained in the feature space, sparsity is increased. By computer experiments we show that in most cases we can reduce the number of support vectors without deteriorating the generalization ability.
Keywords :
matrix decomposition; support vector machines; Cholesky factorization; independent training data; reduced empirical feature space; sparse support vector machines; Constraint optimization; Eigenvalues and eigenfunctions; Kernel; Least squares methods; Newton method; Sparse matrices; Support vector machine classification; Support vector machines; Symmetric matrices; Training data;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2008. IJCNN 2008. (IEEE World Congress on Computational Intelligence). IEEE International Joint Conference on
Conference_Location :
Hong Kong
ISSN :
1098-7576
Print_ISBN :
978-1-4244-1820-6
Electronic_ISBN :
1098-7576
Type :
conf
DOI :
10.1109/IJCNN.2008.4634131
Filename :
4634131
Link To Document :
بازگشت