DocumentCode :
2564446
Title :
Kullback-Leibler divergence rate between probability distributions on sets of different cardinalities
Author :
Vidyasagar, M.
Author_Institution :
Dept. of Bioeng., Univ. of Texas at Dallas, Richardson, TX, USA
fYear :
2010
fDate :
15-17 Dec. 2010
Firstpage :
948
Lastpage :
953
Abstract :
In this paper we generalize the familiar notion of the Kullback-Leibler divergence between two probability distribitions on a finite set to the case where the two probability distributions are defined on sets of different cardinalities. This is achieved by `dilating´ the distribution on the smaller set to one on the larger set. This idea follows. Then, using the log sum inequality, we give a formula for this divergence that permits all computations to be carried out on the set of smaller cardinality. However, computing the divergence is still in general an intractable problem. Hence we give a greedy algorithm for quickly obtaining an upper bound for this divergence. The ideas given here can be used to study the concept of an optimally aggregated model of a Markov or hidden Markov process. This is done in a companion paper.
Keywords :
Markov processes; entropy; set theory; statistical distributions; Kullback-Leibler divergence rate; Shannon entropy; cardinality; greedy algorithm; hidden Markov process; log sum inequality; probability distribution; Complexity theory; Entropy; Greedy algorithms; Markov processes; Mutual information; Probability distribution; Tin;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Decision and Control (CDC), 2010 49th IEEE Conference on
Conference_Location :
Atlanta, GA
ISSN :
0743-1546
Print_ISBN :
978-1-4244-7745-6
Type :
conf
DOI :
10.1109/CDC.2010.5716982
Filename :
5716982
Link To Document :
بازگشت