DocumentCode :
445899
Title :
Yet faster method to optimize SVR hyperparameters based on minimizing cross-validation error
Author :
Kobayashi, Kenji ; Kitakoshi, Daisuke ; Nakano, Ryohei
Author_Institution :
Nagoya Inst. of Technol., Japan
Volume :
2
fYear :
2005
fDate :
31 July-4 Aug. 2005
Firstpage :
871
Abstract :
The performance of support vector (SV) regression deeply depends on its hyperparameters such as an insensitive zone thickness, a penalty factor, kernel function parameters. A method called MCV-SVR was once proposed, which optimizes SVR hyperparameters λ so that a cross-validation error is minimized. The method iterates two steps until convergence; step 1 optimizes parameters θ under given λ, while step 2 improves λ under given θ. Recently a faster version called the MCV-SVR-light was proposed, which accelerates step 2 by pruning. The present paper yet accelerates step 1 of the MCV-SVR-light by pruning without affecting solution quality. Here the pruning means confining the process to support vectors. Our experiments using three data sets show that the proposed method converged faster than the existing methods while the generalization performance remained comparable.
Keywords :
regression analysis; support vector machines; MCV-SVR-light; cross-validation error; insensitive zone thickness; support vector regression; Acceleration; Convergence; Kernel; Lagrangian functions; Neural networks; Optimization methods; Quadratic programming; Training data; Vectors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2005. IJCNN '05. Proceedings. 2005 IEEE International Joint Conference on
Print_ISBN :
0-7803-9048-2
Type :
conf
DOI :
10.1109/IJCNN.2005.1555967
Filename :
1555967
Link To Document :
بازگشت