DocumentCode :
2478647
Title :
Properties of the k-norm pruning algorithm for decision tree classifiers
Author :
Zhong, Mingyu ; Georgiopoulos, Michael ; Anagnost, Georgios C.
Author_Institution :
Sch. of EECS, Univ. of Central Florida, Orlando, FL
fYear :
2008
fDate :
8-11 Dec. 2008
Firstpage :
1
Lastpage :
4
Abstract :
Pruning is one of the key procedures in training decision tree classifiers. It removes trivial rules from the raw knowledge base built from training examples, in order to avoid over-using noisy, conflicting, or fuzzy inputs, so that the refined model can generalize better with unseen cases. In this paper, we present a number of properties of k-norm pruning, a recently proposed pruning algorithm, which has clear theoretical interpretation. In an earlier paper it was shown that k-norm pruning compares very favorably in terms of accuracy and size with minimal cost-complexity pruning and error based pruning, two of the most cited decision tree pruning methods; it was also shown that k-norm pruning is more efficient, at times orders of magnitude more efficient than minimal cost-complexity pruning and error based pruning. In this paper, we demonstrate the validity of the k-norm properties through a series of theorems, and explain their practical significance.
Keywords :
decision trees; pattern classification; decision tree classifiers; decision tree pruning methods; error based pruning; k-norm pruning algorithm; minimal cost-complexity pruning; Classification tree analysis; Costs; Decision trees; Error analysis; Machine learning; Machine learning algorithms; Postal services; Random variables; Robustness; Uncertainty;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Pattern Recognition, 2008. ICPR 2008. 19th International Conference on
Conference_Location :
Tampa, FL
ISSN :
1051-4651
Print_ISBN :
978-1-4244-2174-9
Electronic_ISBN :
1051-4651
Type :
conf
DOI :
10.1109/ICPR.2008.4761277
Filename :
4761277
Link To Document :
بازگشت