• DocumentCode
    64730
  • Title

    Converting Neural Network Language Models into Back-off Language Models for Efficient Decoding in Automatic Speech Recognition

  • Author

    Arisoy, Ebru ; Chen, S.F. ; Ramabhadran, Bhuvana ; Sethy, Abhinav

  • Author_Institution
    ACCES Dept., IBM T. J. Watson Res. Center, Yorktown Heights, NY, USA
  • Volume
    22
  • Issue
    1
  • fYear
    2014
  • fDate
    Jan. 2014
  • Firstpage
    184
  • Lastpage
    192
  • Abstract
    Neural network language models (NNLMs) have achieved very good performance in large-vocabulary continuous speech recognition (LVCSR) systems. Because decoding with NNLMs is computationally expensive, there is interest in developing methods to approximate NNLMs with simpler language models that are suitable for fast decoding. In this work, we propose an approximate method for converting a feedforward NNLM into a back-off n-gram language model that can be used directly in existing LVCSR decoders. We convert NNLMs of increasing order to pruned back-off language models, using lower-order models to constrain the n-grams allowed in higher-order models. In experiments on Broadcast News data, we find that the resulting back-off models retain the bulk of the gain achieved by NNLMs over conventional n-gram language models, and give accuracy improvements as compared to existing methods for converting NNLMs to back-off models. In addition, the proposed approach can be applied to any type of non-back-off language model to enable efficient decoding.
  • Keywords
    feedforward neural nets; natural languages; speech coding; speech recognition; LVCSR system; automatic speech recognition; back-off n-gram language model; decoding; feedforward NNLM; large-vocabulary continuous speech recognition; neural network language model; Artificial neural networks; Computational modeling; Decoding; History; Speech recognition; Training; Back-off language models; language modeling; neural network language models;
  • fLanguage
    English
  • Journal_Title
    Audio, Speech, and Language Processing, IEEE/ACM Transactions on
  • Publisher
    ieee
  • ISSN
    2329-9290
  • Type

    jour

  • DOI
    10.1109/TASLP.2013.2286919
  • Filename
    6645438