DocumentCode :
1313793
Title :
Generalization Bounds of ERM-Based Learning Processes for Continuous-Time Markov Chains
Author :
Chao Zhang ; Dacheng Tao
Author_Institution :
Sch. of Comput. Eng., Nanyang Technol. Univ., Singapore, Singapore
Volume :
23
Issue :
12
fYear :
2012
Firstpage :
1872
Lastpage :
1883
Abstract :
Many existing results on statistical learning theory are based on the assumption that samples are independently and identically distributed (i.i.d.). However, the assumption of i.i.d. samples is not suitable for practical application to problems in which samples are time dependent. In this paper, we are mainly concerned with the empirical risk minimization (ERM) based learning process for time-dependent samples drawn from a continuous-time Markov chain. This learning process covers many kinds of practical applications, e.g., the prediction for a time series and the estimation of channel state information. Thus, it is significant to study its theoretical properties including the generalization bound, the asymptotic convergence, and the rate of convergence. It is noteworthy that, since samples are time dependent in this learning process, the concerns of this paper cannot (at least straightforwardly) be addressed by existing methods developed under the sample i.i.d. assumption. We first develop a deviation inequality for a sequence of time-dependent samples drawn from a continuous-time Markov chain and present a symmetrization inequality for such a sequence. By using the resultant deviation inequality and symmetrization inequality, we then obtain the generalization bounds of the ERM-based learning process for time-dependent samples drawn from a continuous-time Markov chain. Finally, based on the resultant generalization bounds, we analyze the asymptotic convergence and the rate of convergence of the learning process.
Keywords :
Markov processes; convergence; learning (artificial intelligence); risk analysis; statistical analysis; time series; ERM-based learning process; asymptotic convergence; channel state information estimation; continuous time Markov chain; deviation inequality; empirical risk minimization; generalization bound; rate of convergence; statistical learning theory; symmetrization inequality; time series; Channel state information; Complexity theory; Convergence; Estimation; Markov processes; Zinc; Convergence; Markov chain; deviation inequality; empirical risk minimization; generalization bound; rate of convergence; statistical learning theory;
fLanguage :
English
Journal_Title :
Neural Networks and Learning Systems, IEEE Transactions on
Publisher :
ieee
ISSN :
2162-237X
Type :
jour
DOI :
10.1109/TNNLS.2012.2217987
Filename :
6327676
Link To Document :
بازگشت