DocumentCode :
1401765
Title :
A non-Shannon-type conditional inequality of information quantities
Author :
Zhang, Zhen ; Yeung, Raymond W.
Author_Institution :
Commun. Sci. Inst., Univ. of Southern California, Los Angeles, CA, USA
Volume :
43
Issue :
6
fYear :
1997
fDate :
11/1/1997 12:00:00 AM
Firstpage :
1982
Lastpage :
1986
Abstract :
Given n discrete random variables Ω={X1,…,Xn}, associated with any subset α of {1,2,…,n}, there is a joint entropy H(Xα) where Xα={Xi: i∈α}. This can be viewed as a function defined on 2{1,2,…,n} taking values in [0, +∞). We call this function the entropy function of Ω. The nonnegativity of the joint entropies implies that this function is nonnegative; the nonnegativity of the conditional joint entropies implies that this function is nondecreasing; and the nonnegativity of the conditional mutual information implies that this function is two-alternative. These properties are the so-called basic information inequalities of Shannon´s information measures. An entropy function can be viewed as a 2n -1-dimensional vector where the coordinates are indexed by the subsets of the ground set {1,2,…,n}. As introduced by Yeng (see ibid., vol.43, no.6, p.1923-34, 1997) Γn stands for the cone in IR(2n-1) consisting of all vectors which have all these properties. Let Γn* be the set of all 2n -1-dimensional vectors which correspond to the entropy functions of some sets of n discrete random variables. A fundamental information-theoretic problem is whether or not Γ¯n*=Γn. Here Γ¯n * stands for the closure of the set Γn*. We show that Γ¯n* is a convex cone, Γ2*=Γ2, Γ3*≠Γ3, but Γ¯3 *=Γ3. For four random variables, we have discovered a conditional inequality which is not implied by the basic information inequalities of the same set of random variables. This lends an evidence to the plausible conjecture that Γ¯n*≠Γn for n>3
Keywords :
entropy; functional analysis; random processes; Shannon information measures; conditional joint entropies; conditional mutual information; discrete random variables; entropy function; entropy functions; information inequalities; information quantities; information-theoretic problem; nonShannon-type conditional inequality; nondecreasing function; nonnegative function; two-alternative function; vector; Cramer-Rao bounds; Entropy; Mutual information; Random variables;
fLanguage :
English
Journal_Title :
Information Theory, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9448
Type :
jour
DOI :
10.1109/18.641561
Filename :
641561
Link To Document :
بازگشت