Title :
Improved boosting algorithm through weighted k-nearest neighbors classifier
Author :
Gao, Yunlong ; Pan Jin-yan ; Gao, Feng
Author_Institution :
MOE KLINNS Lab., Xi´´an Jiaotong Univ., Xi´´an, China
Abstract :
AdaBoost is known as an effective method for improving the performance of base classifiers both theoretically and empirically. However, previous studies have shown that AdaBoost is prone to overfitting, especially in noisy domains. On the other hand, the k-nearest neighbors (kNN) rule is one of the oldest and simplest methods for pattern classification, when cleverly combined with prior knowledge, it often yields competitive results. In this paper, an improved boosting algorithm is proposed where AdaBoost and kNN naturally complement each other. First, AdaBoost is run on the training data to capitalize on some statistical regularity in the data. Then, a weighted kNN algorithm designed in this paper is run on the feature space composed of classifiers produced by AdaBoost to achieve competitive results. AdaBoost is then used to enhance the classification accuracy and avoid overfitting by editing the data sets using the weighted kNN algorithm for improving the quality of training data. Experiments performed on ten different UCI data sets show that the new Boosting algorithm almost always achieves considerably better classification accuracy than AdaBoost.
Keywords :
learning (artificial intelligence); pattern classification; statistical analysis; AdaBoost; boosting algorithm; kNN algorithm; pattern classification; statistical regularity; weighted k-nearest neighbors classifier; Boosting; Ionosphere; Sonar; AdaBoost; adaptability; feature space; kNN rules; overfitting;
Conference_Titel :
Computer Science and Information Technology (ICCSIT), 2010 3rd IEEE International Conference on
Conference_Location :
Chengdu
Print_ISBN :
978-1-4244-5537-9
DOI :
10.1109/ICCSIT.2010.5563551