Title :
Convergence and Consistency of Regularized Boosting With Weakly Dependent Observations
Author :
Lozano, Aurelie C. ; Kulkarni, Sanjeev R. ; Schapire, Robert E.
Author_Institution :
IBM T. J. Watson Res. Center, Yorktown Heights, NY, USA
Abstract :
This paper studies the statistical convergence and consistency of regularized boosting methods, where the samples need not be independent and identically distributed but can come from stationary weakly dependent sequences. Consistency is proven for the composite classifiers that result from a regularization achieved by restricting the 1-norm of the base classifiers´ weights. The less restrictive nature of sampling considered here is manifested in the consistency result through a generalized condition on the growth of the regularization parameter. The weaker the sample dependence, the faster the regularization parameter is allowed to grow with increasing sample size. A consistency result is also provided for data-dependent choices of the regularization parameter.
Keywords :
data handling; learning (artificial intelligence); pattern classification; statistical analysis; composite classifiers; machine learning; regularization parameter; regularized boosting methods; statistical convergence; weakly dependent observations; Boosting; Convergence; Cost function; Minimization; Prediction algorithms; Random variables; Training data; Bayes-risk consistency; beta-mixing; boosting; classification; dependent data; empirical processes; memory; non-iid; penalized model selection; regularization;
Journal_Title :
Information Theory, IEEE Transactions on
DOI :
10.1109/TIT.2013.2287726