• DocumentCode
    1389164
  • Title

    Hierarchical Bayesian Language Models for Conversational Speech Recognition

  • Author

    Huang, Songfang ; Renals, Steve

  • Author_Institution
    Centre for Speech Technol. Res., Univ. of Edinburgh, Edinburgh, UK
  • Volume
    18
  • Issue
    8
  • fYear
    2010
  • Firstpage
    1941
  • Lastpage
    1954
  • Abstract
    Traditional n -gram language models are widely used in state-of-the-art large vocabulary speech recognition systems. This simple model suffers from some limitations, such as overfitting of maximum-likelihood estimation and the lack of rich contextual knowledge sources. In this paper, we exploit a hierarchical Bayesian interpretation for language modeling, based on a nonparametric prior called Pitman-Yor process. This offers a principled approach to language model smoothing, embedding the power-law distribution for natural language. Experiments on the recognition of conversational speech in multiparty meetings demonstrate that by using hierarchical Bayesian language models, we are able to achieve significant reductions in perplexity and word error rate.
  • Keywords
    maximum likelihood estimation; smoothing methods; speech recognition; Pitman-Yor process; contextual knowledge sources; conversational speech recognition; hierarchical Bayesian language models; language model smoothing; maximum-likelihood estimation; n-gram language models; power-law distribution; word error rate; Automatic speech recognition; Bayesian methods; Context modeling; Error analysis; Maximum likelihood estimation; Natural languages; Power system modeling; Smoothing methods; Speech recognition; Vocabulary; AMI corpus; conversational speech recognition; hierarchical Bayesian model; language model (LM); meetings; smoothing;
  • fLanguage
    English
  • Journal_Title
    Audio, Speech, and Language Processing, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1558-7916
  • Type

    jour

  • DOI
    10.1109/TASL.2010.2040782
  • Filename
    5393057