• DocumentCode
    834046
  • Title

    Variational Bayes for continuous hidden Markov models and its application to active learning

  • Author

    Ji, Shihao ; Krishnapuram, Balaji ; Carin, Lawrence

  • Author_Institution
    Dept. of Electr. & Comput. Eng., Duke Univ., Durham, NC, USA
  • Volume
    28
  • Issue
    4
  • fYear
    2006
  • fDate
    4/1/2006 12:00:00 AM
  • Firstpage
    522
  • Lastpage
    532
  • Abstract
    In this paper, we present a variational Bayes (VB) framework for learning continuous hidden Markov models (CHMMs), and we examine the VB framework within active learning. Unlike a maximum likelihood or maximum a posteriori training procedure, which yield a point estimate of the CHMM parameters, VB-based training yields an estimate of the full posterior of the model parameters. This is particularly important for small training sets since it gives a measure of confidence in the accuracy of the learned model. This is utilized within the context of active learning, for which we acquire labels for those feature vectors for which knowledge of the associated label would be most informative for reducing model-parameter uncertainty. Three active learning algorithms are considered in this paper: 1) query by committee (QBC), with the goal of selecting data for labeling that minimize the classification variance, 2) a maximum expected information gain method that seeks to label data with the goal of reducing the entropy of the model parameters, and 3) an error-reduction-based procedure that attempts to minimize classification error over the test data. The experimental results are presented for synthetic and measured data. We demonstrate that all of these active learning methods can significantly reduce the amount of required labeling, compared to random selection of samples for labeling.
  • Keywords
    Bayes methods; entropy; hidden Markov models; learning (artificial intelligence); maximum likelihood estimation; active learning; classification error minimization; classification variance minimization; continuous hidden Markov models; error-reduction-based procedure; maximum a posteriori training procedure; maximum expected information gain method; maximum likelihood training procedure; model parameter entropy reduction; model-parameter uncertainty reduction; point estimate; query by committee; variational Bayes; Bayesian methods; Context modeling; Hidden Markov models; Labeling; Maximum likelihood estimation; Parameter estimation; Particle measurements; Signal processing algorithms; Testing; Yield estimation; Variational Bayes (VB); active learning (AL); continuous hidden Markov models (CHMMs); error-reduction-based active learning.; maximum expected information gain (MEIG); query by committee (QBC); Algorithms; Artificial Intelligence; Bayes Theorem; Cluster Analysis; Computer Simulation; Information Storage and Retrieval; Markov Chains; Models, Statistical; Pattern Recognition, Automated;
  • fLanguage
    English
  • Journal_Title
    Pattern Analysis and Machine Intelligence, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    0162-8828
  • Type

    jour

  • DOI
    10.1109/TPAMI.2006.85
  • Filename
    1597110