Title :
A comparison of two optimization techniques for sequence discriminative training of deep neural networks
Author :
Saon, George ; Soltau, Hagen
Author_Institution :
IBM T. J. Watson Res. Center, Yorktown Heights, NY, USA
Abstract :
We compare two optimization methods for lattice-based sequence discriminative training of neural network acoustic models: distributed Hessian-free (DHF) and stochastic gradient descent (SGD). Our findings on two different LVCSR tasks suggest that SGD running on a single GPU machine achieves the best accuracy 2.5 times faster than DHF running on multiple non-GPU machines; however, DHF training achieves a higher accuracy at the end of the optimization. In addition, we present an improved modified forward-backward algorithm for computing lattice-based expected loss functions and gradients that results in a 34% speedup for SGD.
Keywords :
gradient methods; graphics processing units; neural nets; stochastic processes; DHF; GPU machine; LVCSR; SGD; deep neural networks; distributed Hessian-free; forward-backward algorithm; lattice based sequence discriminative training; neural network acoustic models; optimization methods; sequence discriminative training; stochastic gradient descent; two optimization technique comparison; Acoustics; Graphics processing units; Hidden Markov models; Lattices; Neural networks; Optimization; Training; distributed Hessian-free optimization; neural network acoustic models; sequence discriminative training; stochastic gradient descent;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on
Conference_Location :
Florence
DOI :
10.1109/ICASSP.2014.6854668