Title :
Yet faster method to optimize SVR hyperparameters based on minimizing cross-validation error
Author :
Kobayashi, Kenji ; Kitakoshi, Daisuke ; Nakano, Ryohei
Author_Institution :
Nagoya Inst. of Technol., Japan
fDate :
31 July-4 Aug. 2005
Abstract :
The performance of support vector (SV) regression deeply depends on its hyperparameters such as an insensitive zone thickness, a penalty factor, kernel function parameters. A method called MCV-SVR was once proposed, which optimizes SVR hyperparameters λ so that a cross-validation error is minimized. The method iterates two steps until convergence; step 1 optimizes parameters θ under given λ, while step 2 improves λ under given θ. Recently a faster version called the MCV-SVR-light was proposed, which accelerates step 2 by pruning. The present paper yet accelerates step 1 of the MCV-SVR-light by pruning without affecting solution quality. Here the pruning means confining the process to support vectors. Our experiments using three data sets show that the proposed method converged faster than the existing methods while the generalization performance remained comparable.
Keywords :
regression analysis; support vector machines; MCV-SVR-light; cross-validation error; insensitive zone thickness; support vector regression; Acceleration; Convergence; Kernel; Lagrangian functions; Neural networks; Optimization methods; Quadratic programming; Training data; Vectors;
Conference_Titel :
Neural Networks, 2005. IJCNN '05. Proceedings. 2005 IEEE International Joint Conference on
Print_ISBN :
0-7803-9048-2
DOI :
10.1109/IJCNN.2005.1555967