DocumentCode :
1204625
Title :
Consistency of support vector machines and other regularized kernel classifiers
Author :
Steinwart, Ingo
Author_Institution :
Los Alamos Nat. Lab., NM
Volume :
51
Issue :
1
fYear :
2005
Firstpage :
128
Lastpage :
142
Abstract :
It is shown that various classifiers that are based on minimization of a regularized risk are universally consistent, i.e., they can asymptotically learn in every classification task. The role of the loss functions used in these algorithms is considered in detail. As an application of our general framework, several types of support vector machines (SVMs) as well as regularization networks are treated. Our methods combine techniques from stochastics, approximation theory, and functional analysis
Keywords :
approximation theory; functional analysis; learning (artificial intelligence); minimisation; pattern classification; stochastic processes; support vector machines; SVM; approximation theory; asymptotic learning; functional analysis; kernel classifier regularization; minimization; regularization network; stochastic; support vector machine; universal consistency; Approximation methods; Fasteners; Functional analysis; Kernel; Machine learning; Pattern recognition; Statistical distributions; Stochastic processes; Support vector machine classification; Support vector machines; Computational learning theory; kernel methods; pattern recognition; regularization; support vector machines (SVMs); universal consistency;
fLanguage :
English
Journal_Title :
Information Theory, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9448
Type :
jour
DOI :
10.1109/TIT.2004.839514
Filename :
1377497
Link To Document :
بازگشت