DocumentCode :
1209231
Title :
Classifiability-based omnivariate decision trees
Author :
Li, Yuanhong ; Dong, Ming ; Kothari, Ravi
Author_Institution :
Dept. of Comput. Sci., Wayne State Univ., Detroit, MI, USA
Volume :
16
Issue :
6
fYear :
2005
Firstpage :
1547
Lastpage :
1560
Abstract :
Top-down induction of decision trees is a simple and powerful method of pattern classification. In a decision tree, each node partitions the available patterns into two or more sets. New nodes are created to handle each of the resulting partitions and the process continues. A node is considered terminal if it satisfies some stopping criteria (for example, purity, i.e., all patterns at the node are from a single class). Decision trees may be univariate, linear multivariate, or nonlinear multivariate depending on whether a single attribute, a linear function of all the attributes, or a nonlinear function of all the attributes is used for the partitioning at each node of the decision tree. Though nonlinear multivariate decision trees are the most powerful, they are more susceptible to the risks of overfitting. In this paper, we propose to perform model selection at each decision node to build omnivariate decision trees. The model selection is done using a novel classifiability measure that captures the possible sources of misclassification with relative ease and is able to accurately reflect the complexity of the subproblem at each node. The proposed approach is fast and does not suffer from as high a computational burden as that incurred by typical model selection algorithms. Empirical results over 26 data sets indicate that our approach is faster and achieves better classification accuracy compared to statistical model select algorithms.
Keywords :
computational complexity; decision diagrams; decision trees; model-based reasoning; pattern classification; set theory; Bayes error; classifiability-based omnivariate decision trees; classification accuracy; data complexity; data density; decision boundary; decision node; linear function; linear multivariate; misclassification; model selection algorithms; node partitioning; node partitions; nonlinear function; nonlinear multivariate decision trees; pattern classification; statistical model select algorithms; univariate; Classification tree analysis; Computer science; Decision trees; Laboratories; Machine vision; Neural networks; Partitioning algorithms; Pattern classification; Pattern recognition; Power generation; Bayes error; data complexity; data density; decision boundary; omnivariate decision trees; Algorithms; Artificial Intelligence; Computer Simulation; Decision Making, Computer-Assisted; Decision Support Techniques; Models, Theoretical; Pattern Recognition, Automated;
fLanguage :
English
Journal_Title :
Neural Networks, IEEE Transactions on
Publisher :
ieee
ISSN :
1045-9227
Type :
jour
DOI :
10.1109/TNN.2005.852864
Filename :
1528531
Link To Document :
بازگشت