• DocumentCode
    1180596
  • Title

    Space-alternating generalized expectation-maximization algorithm

  • Author

    Fessler, Jeffrey A. ; Hero, Alfred O.

  • Author_Institution
    Div. of Nucl. Med., Michigan Univ., Ann Arbor, MI, USA
  • Volume
    42
  • Issue
    10
  • fYear
    1994
  • fDate
    10/1/1994 12:00:00 AM
  • Firstpage
    2664
  • Lastpage
    2677
  • Abstract
    The expectation-maximization (EM) method can facilitate maximizing likelihood functions that arise in statistical estimation problems. In the classical EM paradigm, one iteratively maximizes the conditional log-likelihood of a single unobservable complete data space, rather than maximizing the intractable likelihood function for the measured or incomplete data. EM algorithms update all parameters simultaneously, which has two drawbacks: 1) slow convergence, and 2) difficult maximization steps due to coupling when smoothness penalties are used. The paper describes the space-alternating generalized EM (SAGE) method, which updates the parameters sequentially by alternating between several small hidden-data spaces defined by the algorithm designer. The authors prove that the sequence of estimates monotonically increases the penalized-likelihood objective, derive asymptotic convergence rates, and provide sufficient conditions for monotone convergence in norm. Two signal processing applications illustrate the method: estimation of superimposed signals in Gaussian noise, and image reconstruction from Poisson measurements. In both applications, the SAGE algorithms easily accommodate smoothness penalties and converge faster than the EM algorithms
  • Keywords
    convergence of numerical methods; image reconstruction; iterative methods; maximum likelihood estimation; parameter estimation; random noise; signal processing; statistical analysis; stochastic processes; Gaussian noise; Poisson measurements; SAGE method; asymptotic convergence rates; classical EM paradigm; conditional log-likelihood; image reconstruction; intractable likelihood function; maximization steps; monotone convergence; penalized-likelihood objective; sequence of estimates; signal processing applications; small hidden-data spaces; smoothness penalties; space-alternating generalized expectation-maximization algorithm; statistical estimation problems; sufficient conditions; superimposed signals; unobservable complete data space; Algorithm design and analysis; Convergence; Expectation-maximization algorithms; Extraterrestrial measurements; Gaussian noise; Image reconstruction; Iterative algorithms; Signal processing; Signal processing algorithms; Sufficient conditions;
  • fLanguage
    English
  • Journal_Title
    Signal Processing, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1053-587X
  • Type

    jour

  • DOI
    10.1109/78.324732
  • Filename
    324732