DocumentCode
730846
Title
Recurrent neural network language model training with noise contrastive estimation for speech recognition
Author
Chen, X. ; Liu, X. ; Gales, M.J.F. ; Woodland, P.C.
Author_Institution
Eng. Dept., Univ. of Cambridge, Cambridge, UK
fYear
2015
fDate
19-24 April 2015
Firstpage
5411
Lastpage
5415
Abstract
In recent years recurrent neural network language models (RNNLMs) have been successfully applied to a range of tasks including speech recognition. However, an important issue that limits the quantity of data used, and their possible application areas, is the computational cost in training. A signi??cant part of this cost is associated with the softmax function at the output layer, as this requires a normalization term to be explicitly calculated. This impacts both the training and testing speed, especially when a large output vocabulary is used. To address this problem, noise contrastive estimation (NCE) is explored in RNNLM training. NCE does not require the above normalization during both training and testing. It is insensitive to the output layer size. On a large vocabulary conversational telephone speech recognition task, a doubling in training speed on a GPU and a 56 times speed up in test time evaluation on a CPU were obtained.
Keywords
graphics processing units; language translation; learning (artificial intelligence); recurrent neural nets; speech recognition; vocabulary; CPU; GPU; NCE; RNNLM training; computational cost; large vocabulary conversational telephone speech recognition task; noise contrastive estimation; normalization term; recurrent neural network language model training; softmax function; History; Noise; Recurrent neural networks; Speech recognition; Testing; Training; Vocabulary; GPU; language model; noise contrastive estimation; recurrent neural network; speech recognition;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location
South Brisbane, QLD
Type
conf
DOI
10.1109/ICASSP.2015.7179005
Filename
7179005
Link To Document