DocumentCode :
179593
Title :
Asynchronous stochastic optimization for sequence training of deep neural networks
Author :
Heigold, Georg ; McDermott, Erik ; Vanhoucke, V. ; Senior, Alan ; Bacchiani, Michiel
Author_Institution :
Google Inc., Mountain View, CA, USA
fYear :
2014
fDate :
4-9 May 2014
Firstpage :
5587
Lastpage :
5591
Abstract :
This paper explores asynchronous stochastic optimization for sequence training of deep neural networks. Sequence training requires more computation than frame-level training using pre-computed frame data. This leads to several complications for stochastic optimization, arising from significant asynchrony in model updates under massive parallelization, and limited data shuffling due to utterance-chunked processing. We analyze the impact of these two issues on the efficiency and performance of sequence training. In particular, we suggest a framework to formalize the reasoning about the asynchrony and present experimental results on both small and large scale Voice Search tasks to validate the effectiveness and efficiency of asynchronous stochastic optimization.
Keywords :
neural nets; speech processing; speech recognition; stochastic programming; asynchronous stochastic optimization; deep neural networks; frame-level training; large scale voice search tasks; limited data shuffling; massive parallelization; pre-computed frame data; sequence training; small scale voice search tasks; speech recognition; utterance-chunked processing; Acoustics; Computational modeling; Hidden Markov models; Neural networks; Optimization; Speech; Training; acoustic modeling; asynchronous stochastic optimization; neural networks; sequence training; speech recognition;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on
Conference_Location :
Florence
Type :
conf
DOI :
10.1109/ICASSP.2014.6854672
Filename :
6854672
Link To Document :
بازگشت