DocumentCode :
1374477
Title :
Fast Computation of the Kullback–Leibler Divergence and Exact Fisher Information for the First-Order Moving Average Model
Author :
Makalic, Enes ; Schmidt, Daniel F.
Author_Institution :
Centre for MEGA Epidemiology, Univ. of Melbourne, Carlton, VIC, Australia
Volume :
17
Issue :
4
fYear :
2010
fDate :
4/1/2010 12:00:00 AM
Firstpage :
391
Lastpage :
393
Abstract :
In this note expressions are derived that allow computation of the Kullback-Leibler (K-L) divergence between two first-order Gaussian moving average models in O n(1) time as the sample size n ?? ??. These expressions can also be used to evaluate the exact Fisher information matrix in On(1) time, and provide a basis for an asymptotic expression of the K-L divergence.
Keywords :
Gaussian processes; information theory; moving average processes; Kullback-Leibler divergence; first-order Gaussian moving average models; fisher information; Fisher information; Kullback–Leibler divergence; moving average models;
fLanguage :
English
Journal_Title :
Signal Processing Letters, IEEE
Publisher :
ieee
ISSN :
1070-9908
Type :
jour
DOI :
10.1109/LSP.2009.2039659
Filename :
5371931
Link To Document :
بازگشت