DocumentCode :
134204
Title :
Improving training time of deep neural networkwith asynchronous averaged stochastic gradient descent
Author :
Zhao You ; Bo Xu
Author_Institution :
Interactive Digital Media Technol. Res. Center, Inst. of Autom., Beijing, China
fYear :
2014
fDate :
12-14 Sept. 2014
Firstpage :
446
Lastpage :
449
Abstract :
Deep neural network acoustic models have shown large improvement in performance over Gaussian mixture models (GMMs) in recent studies. Typically, stochastic gradient descent (SGD) is the most popular method for training deep neural networks. However, training DNN with minibatch based SGD is very slow. Because it requires frequent serial training and scanning the whole training set many passes before reaching the asymptotic region, making it difficult to scale to large dataset. Commonly, we can reduce training time from two aspects, reducing the epochs of training and exploring the distributed training algorithm. There are some distributed training algorithms, such as LBFGS, Hessian-free optimization and asynchronous SGD, have proven significantly reducing the training time. In order to further reduce the training time, we attempted to explore training algorithm with fast convergence and combined it with distributed training algorithm. Averaged stochastic gradient descent (ASGD) is proved simple and effective for one pass on-line learning. This paper investigates the asynchronous ASGD algorithm for deep neural network training. We tested asynchronous ASGD on the Mandarin Chinese recorded speech recognition task using deep neural networks. Experimental results show that the performance of one pass asynchronous ASGD is very close to that of multiple passes asynchronous SGD. Meanwhile, we can reduce the training time by a factor of 6.3.
Keywords :
gradient methods; learning (artificial intelligence); neural nets; GMM; Gaussian mixture model; Hessian-free optimization algorithm; LBFGS algorithm; Mandarin Chinese recorded speech recognition task; SGD method; asynchronous SGD algorithm; asynchronous averaged stochastic gradient descent; deep neural network acoustic models; minibatch based SGD; network training time; one pass online learning; Acoustics; Neural networks; Optimization; Speech; Speech recognition; Stochastic processes; Training; Asynchronous averaged SGD; deep neural network; one pass learning; speech recognition;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Chinese Spoken Language Processing (ISCSLP), 2014 9th International Symposium on
Conference_Location :
Singapore
Type :
conf
DOI :
10.1109/ISCSLP.2014.6936596
Filename :
6936596
Link To Document :
بازگشت