DocumentCode :
1425839
Title :
Mutual Information, Relative Entropy, and Estimation in the Poisson Channel
Author :
Atar, Rami ; Weissman, Tsachy
Author_Institution :
Israel Inst. of Technol., Technion - Israel Inst. of Technol., Haifa, Israel
Volume :
58
Issue :
3
fYear :
2012
fDate :
3/1/2012 12:00:00 AM
Firstpage :
1302
Lastpage :
1318
Abstract :
Let be a nonnegative random variable and let the conditional distribution of a random variable , given , be Poisson , for a parameter . We identify a natural loss function such that: (1) the derivative of the mutual information between and with respect to is equal to the minimum mean loss in estimating based on , regardless of the distribution of ; (2) when is estimated based on by a mismatched estimator that would have minimized the expected loss had , the integral over all values of of the excess mean loss is equal to the relative entropy between and . For a continuous time setting where is a nonnegative stochastic process and the conditional law of , given , is that of a non-homogeneous Poisson process with intensity function , under the same loss function: (1) the minimum mean loss in causal filtering when is equal to the expected value of the minimum mean loss in noncausal filtering (smoothing) achieved with a channel whose parameter is uniformly distributed between 0 and . Bridging the two quantities is the mutual information between and ; (2) this relationship between the mean losses in causal and noncausal filtering holds also in the case where the filters employed are mismatched, i.e., optimized assuming a law on which is not the true one. Bridging the two quantities in this case is the sum of the mutual information and the relative entropy between the true and the mismatched distribution of . Thus, relative entropy quantifies the excess estimation loss due to mismatch in this setting. These results are parallel to those recently found for the Gaussian channel: the I-MMSE relationship of Guo , the relative entropy and mismatched estimation relationship of Verdú, and the relationship between causal and noncasual mismatched estimation of Weissman.
Keywords :
Gaussian channels; Poisson distribution; channel bank filters; channel estimation; stochastic processes; Gaussian channel; I-MMSE relationship; Poisson channel estimation loss; causal filtering; conditional distribution; intensity function; minimum mean loss; mismatched distribution; mismatched estimator; natural loss function; noncasual mismatched estimation relationship; noncausal filtering; nonhomogeneous Poisson process; nonnegative random variable; nonnegative stochastic process; relative entropy; Channel estimation; Entropy; Estimation; Loss measurement; Mutual information; Random variables; Signal to noise ratio; Causal estimation; Girsanov transformation; I-MMSE; Poisson channel; Shannon theory; divergence; mismatched estimation; mutual information; nonlinear filtering; point process; relative entropy; statistics;
fLanguage :
English
Journal_Title :
Information Theory, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9448
Type :
jour
DOI :
10.1109/TIT.2011.2172572
Filename :
6134671
Link To Document :
بازگشت