DocumentCode :
2373556
Title :
Leveraging Κ-nn for generic classification boosting
Author :
Piro, Paolo ; Noch, Richard ; Nielsen, Frank ; Barlaud, Michel
Author_Institution :
CNRS, Univ. of Nice-Sophia Antipolis, Sophia Antipolis, France
fYear :
2010
fDate :
Aug. 29 2010-Sept. 1 2010
Firstpage :
142
Lastpage :
147
Abstract :
k-nearest neighbors (k-NN) voting rules are an effective tool in countless many machine learning techniques. In spite of its simplicity, k-NN classification is very attractive to practitioners, as it has shown very good performances in practical applications. However, it suffers from various drawbacks, like sensitivity to “noisy” prototypes and poor generalization properties when dealing with sparse, high-dimensional data. In this paper, we tackle the k-NN classification problem at its core by providing a novel k-NN boosting approach. We propose a Universal Nearest Neighbors (UNN) algorithm, which induces a leveraged k-NN rule by globally minimizing a surrogate risk upper bounding the empirical misclassification rate over training data. Interestingly, this surrogate risk can be arbitrary chosen from a class of Bregman loss functions, including the familiar exponential, logistic and squared losses. Furthermore, we show that UNN allows to efficiently filter a dataset of instances by keeping only a small fraction of data. Experimental results on the synthetic Ripley´s dataset show that such a filtering strategy is able to reject “noisy” examples, and yields a classification error close to the optimal Bayes error. Experiments on standard UCI datasets show significant improvements over the current state of the art.
Keywords :
learning (artificial intelligence); pattern classification; generic classification boosting; k-nearest neighbors voting rules; machine learning techniques; universal nearest neighbors algorithm; Artificial neural networks; Boosting; Indexes; Measurement; Minimization; Nearest neighbor searches; Training;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Machine Learning for Signal Processing (MLSP), 2010 IEEE International Workshop on
Conference_Location :
Kittila
ISSN :
1551-2541
Print_ISBN :
978-1-4244-7875-0
Electronic_ISBN :
1551-2541
Type :
conf
DOI :
10.1109/MLSP.2010.5589248
Filename :
5589248
Link To Document :
بازگشت