• DocumentCode
    672374
  • Title

    Accelerating recurrent neural network training via two stage classes and parallelization

  • Author

    Zhiheng Huang ; Zweig, Geoffrey ; Levit, Michael ; Dumoulin, Benoit ; Oguz, Barlas ; Shawn Chang

  • Author_Institution
    Speech at Microsoft, Mountain View, CA, USA
  • fYear
    2013
  • fDate
    8-12 Dec. 2013
  • Firstpage
    326
  • Lastpage
    331
  • Abstract
    Recurrent neural network (RNN) language models have proven to be successful to lower the perplexity and word error rate in automatic speech recognition (ASR). However, one challenge to adopt RNN language models is due to their heavy computational cost in training. In this paper, we propose two techniques to accelerate RNN training: 1) two stage class RNN and 2) parallel RNN training. In experiments on Microsoft internal short message dictation (SMD) data set, two stage class RNNs and parallel RNNs not only result in equal or lower WERs compared to original RNNs but also accelerate training by 2 and 10 times respectively. It is worth noting that two stage class RNN speedup can also be applied to test stage, which is essential to reduce the latency in real time ASR applications.
  • Keywords
    learning (artificial intelligence); parallel processing; recurrent neural nets; speech recognition; Microsoft internal short message dictation data set; RNN language models; SMD data set; WER; automatic speech recognition; latency reduction; parallel RNN training; real time ASR applications; recurrent neural network training; two-stage class RNN training; word error rate; Complexity theory; Computational modeling; Data models; Mathematical model; Recurrent neural networks; Training; Training data; hierarchical classes; language modeling; parallelization; recurrent neural network (RNN); speed up;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Automatic Speech Recognition and Understanding (ASRU), 2013 IEEE Workshop on
  • Conference_Location
    Olomouc
  • Type

    conf

  • DOI
    10.1109/ASRU.2013.6707751
  • Filename
    6707751