Title :
Accelerating recurrent neural network training via two stage classes and parallelization
Author :
Zhiheng Huang ; Zweig, Geoffrey ; Levit, Michael ; Dumoulin, Benoit ; Oguz, Barlas ; Shawn Chang
Author_Institution :
Speech at Microsoft, Mountain View, CA, USA
Abstract :
Recurrent neural network (RNN) language models have proven to be successful to lower the perplexity and word error rate in automatic speech recognition (ASR). However, one challenge to adopt RNN language models is due to their heavy computational cost in training. In this paper, we propose two techniques to accelerate RNN training: 1) two stage class RNN and 2) parallel RNN training. In experiments on Microsoft internal short message dictation (SMD) data set, two stage class RNNs and parallel RNNs not only result in equal or lower WERs compared to original RNNs but also accelerate training by 2 and 10 times respectively. It is worth noting that two stage class RNN speedup can also be applied to test stage, which is essential to reduce the latency in real time ASR applications.
Keywords :
learning (artificial intelligence); parallel processing; recurrent neural nets; speech recognition; Microsoft internal short message dictation data set; RNN language models; SMD data set; WER; automatic speech recognition; latency reduction; parallel RNN training; real time ASR applications; recurrent neural network training; two-stage class RNN training; word error rate; Complexity theory; Computational modeling; Data models; Mathematical model; Recurrent neural networks; Training; Training data; hierarchical classes; language modeling; parallelization; recurrent neural network (RNN); speed up;
Conference_Titel :
Automatic Speech Recognition and Understanding (ASRU), 2013 IEEE Workshop on
Conference_Location :
Olomouc
DOI :
10.1109/ASRU.2013.6707751