• DocumentCode
    16520
  • Title

    VC-Dimension of Univariate Decision Trees

  • Author

    Yildiz, Olcay Taner

  • Author_Institution
    Dept. of Comput. Eng., Isik Univ., İstanbul, Turkey
  • Volume
    26
  • Issue
    2
  • fYear
    2015
  • fDate
    Feb. 2015
  • Firstpage
    378
  • Lastpage
    387
  • Abstract
    In this paper, we give and prove the lower bounds of the Vapnik-Chervonenkis (VC)-dimension of the univariate decision tree hypothesis class. The VC-dimension of the univariate decision tree depends on the VC-dimension values of its subtrees and the number of inputs. Via a search algorithm that calculates the VC-dimension of univariate decision trees exhaustively, we show that our VC-dimension bounds are tight for simple trees. To verify that the VC-dimension bounds are useful, we also use them to get VC-generalization bounds for complexity control using structural risk minimization in decision trees, i.e., pruning. Our simulation results show that structural risk minimization pruning using the VC-dimension bounds finds trees that are more accurate as those pruned using cross validation.
  • Keywords
    computational complexity; decision trees; learning (artificial intelligence); search problems; statistical analysis; VC-dimension; VC-generalization bounds; Vapnik-Chervonenkis dimension; complexity control; cross validation; search algorithm; statistical learning theory; structural risk minimization; subtrees; univariate decision tree hypothesis class; Complexity theory; Decision trees; Labeling; Learning systems; Risk management; Training; Upper bound; Computation theory; Vapnik--Chervonenkis (VC)-dimension.; Vapnik???Chervonenkis (VC)-dimension; decision trees; learning; machine learning; supervised learning;
  • fLanguage
    English
  • Journal_Title
    Neural Networks and Learning Systems, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    2162-237X
  • Type

    jour

  • DOI
    10.1109/TNNLS.2014.2385837
  • Filename
    7008521