• DocumentCode
    3433172
  • Title

    Discriminative method for recurrent neural network language models

  • Author

    Tachioka, Yuuki ; Watanabe, Shinji

  • Author_Institution
    Inf. Technol. R&D Center, Mitsubishi Electr. Corp., Kamakura, Japan
  • fYear
    2015
  • fDate
    19-24 April 2015
  • Firstpage
    5386
  • Lastpage
    5390
  • Abstract
    A recurrent neural network language model (RNN-LM) can use a long word context more than can an n-gram language model, and its effective has recently been shown in its accomplishment of automatic speech recognition (ASR) tasks. However, the training criteria of RNN-LM are based on cross entropy (CE) between predicted and reference words. In addition, unlike the discriminative training of acoustic models and discriminative language models (DLM), these criteria do not explicitly consider discriminative criteria calculated from ASR hypotheses and references. This paper proposes a discriminative training method for RNN-LM by additionally considering a discriminative criterion to CE. We use the log-likelihood ratio of the ASR hypotheses and references as an discriminative criterion. The proposed training criterion emphasizes the effect of improperly recognized words relatively compared to the effect of correct words, which are discounted in training. Experiments on a large vocabulary continuous speech recognition task show that our proposed method improves the RNN-LM baseline. In addition, combining the proposed discriminative RNN-LM and DLM further shows its effectiveness.
  • Keywords
    acoustic signal processing; entropy; natural language processing; recurrent neural nets; speech recognition; ASR; DLM; RNN-LM; acoustic models; automatic speech recognition; continuous speech recognition task; cross entropy; discriminative language models; discriminative training method; n-gram language model; recurrent neural network language models; word context; Artificial neural networks; Irrigation; Laboratories; Mathematical model; Training; Yttrium; Speech recognition; discriminative criterion; language model; log-likelihood ratio; recurrent neural network;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
  • Conference_Location
    South Brisbane, QLD
  • Type

    conf

  • DOI
    10.1109/ICASSP.2015.7179000
  • Filename
    7179000