Title :
Sparse approximation using least squares support vector machines
Author :
Suykens, J.A.K. ; Lukas, L. ; Vandewalle, J.
Author_Institution :
ESAT, Katholieke Univ., Leuven, Heverlee, Belgium
Abstract :
In least squares support vector machines (LS-SVMs) for function estimation Vapnik´s ε-insensitive loss function has been replaced by a cost function which corresponds to a form of ridge regression. In this way nonlinear function estimation is done by solving a linear set of equations instead of solving a quadratic programming problem. The LS-SVM formulation also involves less tuning parameters. However, a drawback is that sparseness is lost in the LS-SVM case. In this paper we investigate imposing sparseness by pruning support values from the sorted support value spectrum which results from the solution to the linear system
Keywords :
least squares approximations; nonlinear functions; radial basis function networks; sparse matrices; statistical analysis; cost function; least squares support vector machines; linear equations set; nonlinear function estimation; ridge regression; sorted support value spectrum; sparse approximation; sparseness; tuning parameters; Cost function; Ear; Equations; Kernel; Least squares approximation; Least squares methods; Linear systems; Quadratic programming; Support vector machine classification; Support vector machines;
Conference_Titel :
Circuits and Systems, 2000. Proceedings. ISCAS 2000 Geneva. The 2000 IEEE International Symposium on
Conference_Location :
Geneva
Print_ISBN :
0-7803-5482-6
DOI :
10.1109/ISCAS.2000.856439