DocumentCode
672374
Title
Accelerating recurrent neural network training via two stage classes and parallelization
Author
Zhiheng Huang ; Zweig, Geoffrey ; Levit, Michael ; Dumoulin, Benoit ; Oguz, Barlas ; Shawn Chang
Author_Institution
Speech at Microsoft, Mountain View, CA, USA
fYear
2013
fDate
8-12 Dec. 2013
Firstpage
326
Lastpage
331
Abstract
Recurrent neural network (RNN) language models have proven to be successful to lower the perplexity and word error rate in automatic speech recognition (ASR). However, one challenge to adopt RNN language models is due to their heavy computational cost in training. In this paper, we propose two techniques to accelerate RNN training: 1) two stage class RNN and 2) parallel RNN training. In experiments on Microsoft internal short message dictation (SMD) data set, two stage class RNNs and parallel RNNs not only result in equal or lower WERs compared to original RNNs but also accelerate training by 2 and 10 times respectively. It is worth noting that two stage class RNN speedup can also be applied to test stage, which is essential to reduce the latency in real time ASR applications.
Keywords
learning (artificial intelligence); parallel processing; recurrent neural nets; speech recognition; Microsoft internal short message dictation data set; RNN language models; SMD data set; WER; automatic speech recognition; latency reduction; parallel RNN training; real time ASR applications; recurrent neural network training; two-stage class RNN training; word error rate; Complexity theory; Computational modeling; Data models; Mathematical model; Recurrent neural networks; Training; Training data; hierarchical classes; language modeling; parallelization; recurrent neural network (RNN); speed up;
fLanguage
English
Publisher
ieee
Conference_Titel
Automatic Speech Recognition and Understanding (ASRU), 2013 IEEE Workshop on
Conference_Location
Olomouc
Type
conf
DOI
10.1109/ASRU.2013.6707751
Filename
6707751
Link To Document