DocumentCode :
259610
Title :
Improved kNN Rule for Small Training Sets
Author :
Cheamanunkul, Sunsern ; Freund, Yoav
Author_Institution :
Dept. of Comput. Sci. & Eng., Univ. of California, San Diego, La Jolla, CA, USA
fYear :
2014
fDate :
3-6 Dec. 2014
Firstpage :
201
Lastpage :
206
Abstract :
The traditional k-NN classification rule predicts a label based on the most common label of the k nearest neighbors (the plurality rule). It is known that the plurality rule is optimal when the number of examples tends to infinity. In this paper we show that the plurality rule is sub-optimal when the number of labels is large and the number of examples is small. We propose a simple k-NN rule that takes into account the labels of all of the neighbors, rather than just the most common label. We present a number of experiments on both synthetic datasets and real-world datasets, including MNIST and SVHN. We show that our new rule can achieve lower error rates compared to the majority rule in many cases.
Keywords :
neural nets; pattern classification; set theory; MNIST; SVHN; error rates; improved kNN rule; k nearest neighbors; k-NN classification rule; optimal plurality rule; real-world datasets; training sets; Computer science; Data models; Educational institutions; Electronic mail; Error analysis; Prediction algorithms; Training;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Machine Learning and Applications (ICMLA), 2014 13th International Conference on
Conference_Location :
Detroit, MI
Type :
conf
DOI :
10.1109/ICMLA.2014.37
Filename :
7033115
Link To Document :
بازگشت