DocumentCode :
1755895
Title :
Probability Functionals for Self-Consistent and Invariant Inference: Entropy and Fisher Information
Author :
Langley, Robin S.
Author_Institution :
Dept. of Eng., Univ. of Cambridge, Cambridge, UK
Volume :
59
Issue :
7
fYear :
2013
fDate :
41456
Firstpage :
4397
Lastpage :
4407
Abstract :
Two existing methods of probabilistic inference are based on variational principles: maximum entropy and minimum Fisher information. In each case, a probability density function is inferred by setting the first variation of a functional to zero, subject to information constraints. This study considers whether other functionals could be used for this purpose, and by starting with requirements for self-consistency and invariance, it is shown that the most general admissible functional is just a linear combination of entropy and Fisher information, with the proviso that the normal definition of Fisher information is modified by the inclusion of a prior. This amounts to an axiomatic derivation of entropy and Fisher information. The concern is with continuous random variables and both the single- and multivariable cases are considered. A number of examples are considered to compare inference based on entropy with that based on Fisher information, and to highlight the role of boundary conditions for inference based on Fisher information.
Keywords :
entropy; probability; variational techniques; admissible functional; invariant inference; maximum entropy; minimum Fisher information; probabilistic inference; probability density function; self-consistent inference; variational principle; Boundary conditions; Entropy; Equations; Probabilistic logic; Probability density function; Random variables; Transforms; Entropy; Fisher information; inference; maximum entropy;
fLanguage :
English
Journal_Title :
Information Theory, IEEE Transactions on
Publisher :
ieee
ISSN :
0018-9448
Type :
jour
DOI :
10.1109/TIT.2013.2252396
Filename :
6478820
Link To Document :
بازگشت