Title :
Corrective Classification: Classifier Ensembling with Corrective and Diverse Base Learners
Author :
Zhang, Yan ; Zhu, Xingquan ; Wu, Xindong
Author_Institution :
Dept. of Comput. Sci., Vermont Univ., Burlington, VT
Abstract :
Empirical studies on supervised learning have shown that ensembling methods lead to a model superior to the one built from a single learner under many circumstances especially when learning from imperfect, such as biased or noise infected, information sources. In this paper, we provide a novel corrective classification (C2) design, which incorporates error detection, data cleansing and Bootstrap sampling to construct base learners that constitute the classifier ensemble. The essential goal is to reduce noise impacts and eventually enhance the learners built from noise corrupted data. We further analyze the importance of both the accuracy and diversity of base learners in ensembling, in order to shed some light on the mechanism under which C2 works. Experimental comparisons will demonstrate that C2 is not only superior to the learner built from the original noisy sources, but also more reliable than bagging or the aggressive classifier ensemble (ACE), which are two degenerate components/variants of C2.
Keywords :
data mining; error detection; learning (artificial intelligence); pattern classification; bootstrap sampling; classifier ensembling; corrective base learners; corrective classification; data cleansing; diverse base learners; error detection; noise impacts; supervised learning; Bagging; Computer science; Data analysis; Data mining; Data preprocessing; Error correction; Noise reduction; Robustness; Sampling methods; Supervised learning;
Conference_Titel :
Data Mining, 2006. ICDM '06. Sixth International Conference on
Conference_Location :
Hong Kong
Print_ISBN :
0-7695-2701-7
DOI :
10.1109/ICDM.2006.45