Title :
Scalable Training of Sparse Linear SVMs
Author :
Guo-Xun Yuan ; Kwan-Liu Ma
Author_Institution :
Comput. Sci. Dept., Univ. of California, Davis, Davis, CA, USA
Abstract :
Sparse linear support vector machines have been widely applied to variable selection in many applications. For large data, managing the cost of training a sparse model with good predication performance is an essential topic. In this work, we propose a scalable training algorithm for large-scale data with millions of examples and features. We develop a dual alternating direction method for solving L1-regularized linear SVMs. The learning procedure simply involves quadratic programming in the same form as the standard SVM dual, followed by a soft-thresholding operation. The proposed training algorithm possesses two favorable properties. First, it is a decomposable algorithm by which a large problem can be reduced to small ones. Second, the sparsity of intermediate solutions is maintained throughout the training process. It naturally promotes the solution sparsity by soft-thresholding. We demonstrate that, by experiments, our method outperforms state-of-the-art approaches on large-scale benchmark data sets. We also show that it is well suited for training large sparse models on a distributed system.
Keywords :
pattern classification; quadratic programming; support vector machines; L1-regularized linear SVM; decomposable algorithm; distributed system; dual alternating direction method; large-scale data; learning procedure; predication performance; quadratic programming; scalable training; soft-thresholding operation; sparse classifier; sparse linear SVM; sparse linear support vector machine; variable selection; Approximation methods; Convergence; Data models; Minimization; Standards; Support vector machines; Training; L1 regularization; large-scale linear classification; optimization techniques;
Conference_Titel :
Data Mining (ICDM), 2012 IEEE 12th International Conference on
Conference_Location :
Brussels
Print_ISBN :
978-1-4673-4649-8
DOI :
10.1109/ICDM.2012.157