DocumentCode :
847151
Title :
On competitive prediction and its relation to rate-distortion theory
Author :
Weissman, Tsachy ; Merhav, Neri
Author_Institution :
Dept. of Electr. Eng., Stanford Univ., CA, USA
Volume :
49
Issue :
12
fYear :
2003
Firstpage :
3185
Lastpage :
3194
Abstract :
Consider the normalized cumulative loss of a predictor F on the sequence xn=(x1,...,xn), denoted LF(xn). For a set of predictors G, let L(G,xn)=minF∈GLF(xn) denote the loss of the best predictor in the class on xn. Given the stochastic process X=X1,X2,..., we look at EL(G,Xn), termed the competitive predictability of G on Xn. Our interest is in the optimal predictor set of size M, i.e., the predictor set achieving min|G|≤MEL(G,Xn). When M is subexponential in n, simple arguments show that min|G|≤MEL(G,Xn) coincides, for large n, with the Bayesian envelope minFELF(Xn). We investigate the behavior, for large n, of min|G|≤enREL(G,Xn), which we term the competitive predictability of X at rate R. We show that whenever X has an autoregressive representation via a predictor with an associated independent and identically distributed (i.i.d.) innovation process, its competitive predictability is given by the distortion-rate function of that innovation process. Indeed, it will be argued that by viewing G as a rate-distortion codebook and the predictors in it as codewords allowed to base the reconstruction of each symbol on the past unquantized symbols, the result can be considered as the source-coding analog of Shannon´s classical result that feedback does not increase the capacity of a memoryless channel. For a general process X, we show that the competitive predictability is lower-bounded by the Shannon lower bound (SLB) on the distortion-rate function of X and upper-bounded by the distortion-rate function of any (not necessarily memoryless) innovation process through which the process X has an autoregressive representation. Thus, the competitive predictability is also precisely characterized whenever X can be autoregressively represented via an innovation process for which the SLB is tight. The error exponent, i.e., the exponential behavior of min|G|≤exp(nR)Pr(L(G,Xn)>d), is also characterized for processes that can be autoregressively represented with an i.i.d. innovation process.
Keywords :
channel capacity; error statistics; prediction theory; rate distortion theory; redundancy; source coding; stochastic processes; Bayesian envelope; SLB; Shannon classical result; Shannon lower bound; autoregressive representation; competitive prediction; distortion-rate function; error exponent; expert advice; innovation process; memoryless channel capacity; normalized cumulative loss; rate-distortion codebook; rate-distortion theory; redundancy; scandiction; source-coding; stochastic process; strong converse; symbol reconstruction; unquantized symbols; Bayesian methods; Helium; Memoryless systems; Neural networks; Rate distortion theory; Rate-distortion; Redundancy; Stochastic processes; Technological innovation; Uncertainty;
fLanguage :
English
Journal_Title :
Information Theory, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9448
Type :
jour
DOI :
10.1109/TIT.2003.820014
Filename :
1255544
Link To Document :
بازگشت