• DocumentCode
    1297380
  • Title

    Physics of the Shannon Limits

  • Author

    Merhav, Neri

  • Author_Institution
    Dept. of Electr. Eng., Technion - Israel Inst. of Technol., Haifa, Israel
  • Volume
    56
  • Issue
    9
  • fYear
    2010
  • Firstpage
    4274
  • Lastpage
    4285
  • Abstract
    We provide a simple physical interpretation, in the context of the second law of thermodynamics, to the information inequality (a.k.a. the Gibbs inequality, which is also equivalent to the log-sum inequality), asserting that the relative entropy between two probability distributions cannot be negative. Since this inequality stands at the basis of the data processing theorem (DPT), and the DPT in turn is at the heart of most, if not all, proofs of converse theorems in Shannon theory, it is observed that conceptually, the roots of fundamental limits of information theory can actually be attributed to the laws of physics, in particular, the second law of thermodynamics, and indirectly, also the law of energy conservation. By the same token, in the other direction: one can view the second law as stemming from information-theoretic principles.
  • Keywords
    energy conservation; entropy; statistical distributions; Gibbs inequality; Shannon limits; data processing theorem; energy conservation; information inequality; information theory; log-sum inequality; physical interpretation; probability distribution; relative entropy; second law of thermodynamics; Cramer-Rao bounds; Data processing; Energy conservation; Entropy; Heart; Information theory; Markov processes; Mutual information; Physics; Probability distribution; Thermodynamics; Data processing theorem; Gibbs´ inequality; divergence; entropy; mutual information; relative entropy; second law of thermodynamics;
  • fLanguage
    English
  • Journal_Title
    Information Theory, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0018-9448
  • Type

    jour

  • DOI
    10.1109/TIT.2010.2053867
  • Filename
    5550388