DocumentCode :
680730
Title :
Should the Same Learners Be Used Both within Wrapper Feature Selection and for Building Classification Models?
Author :
Wald, Randall ; Khoshgoftaar, Taghi M. ; Napolitano, Antonio
Author_Institution :
Florida Atlantic Univ., Boca Raton, FL, USA
fYear :
2013
fDate :
4-6 Nov. 2013
Firstpage :
439
Lastpage :
445
Abstract :
Due to the problem of high-dimensionality(datasets which contain many independent attributes or features), feature selection has become an important part of data mining research. One popular form of feature selection, wrapper selection, chooses the best features by directly addressing the question of which features build the best models. Various feature subsets are used to build classification models, and the performance of these models is the score of each feature subset. The feature subset with the best score is then used to build the final classification model. As wrappers use a classification algorithm (learner) both to select the features and to build a predictive model, it has been traditional to use the same learner for both, such that the features chosen will be those which optimize that model´s performance. However, no research has considered whether having different learners operate inside and outside the wrapper (that is, for selectingthe features and for building the final model) might actually result in improved classification performance. In this work, weconsider five learners both inside the wrapper and for building the classification model, along with two datasets drawn fromthe domain of Twitter profile mining. By considering both the raw performance values and a statistical analysis, we find that contrary to intuition, usually the best performance for a given choice of external learner is not found by using the same learner within the wrapper. Instead, the Naïve Bayes learner is usually the best choice for selecting features, regardless of which learner is used for the external model. We also find that Multi-Layer Perceptron is able to build consistent classification models for many different choices of internal learner. Finally, the 5-Nearest Neighbor learner gave poor results both inside and outside the wrapper.
Keywords :
belief networks; data mining; feature selection; pattern classification; statistical analysis; 5-nearest neighbor learner; Naïve Bayes learner; Twitter profile mining; classification models; internal learner; multilayer perceptron; statistical analysis; wrapper feature selection; Buildings; Computational modeling; Feature extraction; Measurement; Niobium; Support vector machines; Twitter; Twitter; Wrapper feature selection; classification algorithm; learner; social bots;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Tools with Artificial Intelligence (ICTAI), 2013 IEEE 25th International Conference on
Conference_Location :
Herndon, VA
ISSN :
1082-3409
Print_ISBN :
978-1-4799-2971-9
Type :
conf
DOI :
10.1109/ICTAI.2013.72
Filename :
6735283
Link To Document :
بازگشت