DocumentCode
9323
Title
Finding Potential Support Vectors in Separable Classification Problems
Author
Varagnolo, Damiano ; Del Favero, Simone ; Dinuzzo, Francesco ; Schenato, L. ; Pillonetto, G.
Author_Institution
Autom. Control Lab., KTH R. Inst. of Technol., Stockholm, Sweden
Volume
24
Issue
11
fYear
2013
fDate
Nov. 2013
Firstpage
1799
Lastpage
1813
Abstract
This paper considers the classification problem using support vector (SV) machines and investigates how to maximally reduce the size of the training set without losing information. Under separable data set assumptions, we derive the exact conditions stating which observations can be discarded without diminishing the overall information content. For this purpose, we introduce the concept of potential SVs, i.e., those data that can become SVs when future data become available. To complement this, we also characterize the set of discardable vectors (DVs), i.e., those data that, given the current data set, can never become SVs. Thus, these vectors are useless for future training purposes and can eventually be removed without loss of information. Then, we provide an efficient algorithm based on linear programming that returns the potential and DVs by constructing a simplex tableau. Finally, we compare it with alternative algorithms available in the literature on some synthetic data as well as on data sets from standard repositories.
Keywords
linear programming; pattern classification; support vector machines; DV; SVM; discardable vectors; information content; linear programming; potential SV concept; separable classification problem; separable data set assumptions; simplex tableau; support vector machines; Data discardability conditions; discardable vectors; linear programming; potential support vectors; separable data sets; support vector machines;
fLanguage
English
Journal_Title
Neural Networks and Learning Systems, IEEE Transactions on
Publisher
ieee
ISSN
2162-237X
Type
jour
DOI
10.1109/TNNLS.2013.2264731
Filename
6547234
Link To Document