DocumentCode :
16520
Title :
VC-Dimension of Univariate Decision Trees
Author :
Yildiz, Olcay Taner
Author_Institution :
Dept. of Comput. Eng., Isik Univ., İstanbul, Turkey
Volume :
26
Issue :
2
fYear :
2015
fDate :
Feb. 2015
Firstpage :
378
Lastpage :
387
Abstract :
In this paper, we give and prove the lower bounds of the Vapnik-Chervonenkis (VC)-dimension of the univariate decision tree hypothesis class. The VC-dimension of the univariate decision tree depends on the VC-dimension values of its subtrees and the number of inputs. Via a search algorithm that calculates the VC-dimension of univariate decision trees exhaustively, we show that our VC-dimension bounds are tight for simple trees. To verify that the VC-dimension bounds are useful, we also use them to get VC-generalization bounds for complexity control using structural risk minimization in decision trees, i.e., pruning. Our simulation results show that structural risk minimization pruning using the VC-dimension bounds finds trees that are more accurate as those pruned using cross validation.
Keywords :
computational complexity; decision trees; learning (artificial intelligence); search problems; statistical analysis; VC-dimension; VC-generalization bounds; Vapnik-Chervonenkis dimension; complexity control; cross validation; search algorithm; statistical learning theory; structural risk minimization; subtrees; univariate decision tree hypothesis class; Complexity theory; Decision trees; Labeling; Learning systems; Risk management; Training; Upper bound; Computation theory; Vapnik--Chervonenkis (VC)-dimension.; Vapnik???Chervonenkis (VC)-dimension; decision trees; learning; machine learning; supervised learning;
fLanguage :
English
Journal_Title :
Neural Networks and Learning Systems, IEEE Transactions on
Publisher :
ieee
ISSN :
2162-237X
Type :
jour
DOI :
10.1109/TNNLS.2014.2385837
Filename :
7008521
Link To Document :
بازگشت