Title :
Efficient lattice rescoring using recurrent neural network language models
Author :
Liu, Xindong ; Wang, Yannan ; Chen, Xia ; Gales, Mark J.F. ; Woodland, Philip C.
Author_Institution :
Eng. Dept., Cambridge Univ., Cambridge, UK
Abstract :
Recurrent neural network language models (RNNLM) have become an increasingly popular choice for state-of-the-art speech recognition systems due to their inherently strong generalization performance. As these models use a vector representation of complete history contexts, RNNLMs are normally used to rescore N-best lists. Motivated by their intrinsic characteristics, two novel lattice rescoring methods for RNNLMs are investigated in this paper. The first uses an n-gram style clustering of history contexts. The second approach directly exploits the distance measure between hidden history vectors. Both methods produced 1-best performance comparable with a 10k-best rescoring baseline RNNLM system on a large vocabulary conversational telephone speech recognition task. Significant lattice size compression of over 70% and consistent improvements after confusion network (CN) decoding were also obtained over the N-best rescoring approach.
Keywords :
recurrent neural nets; speech recognition; RNNLM system; confusion network decoding; distance measure; hidden history vectors; large vocabulary conversational telephone speech recognition task; lattice size compression; n-gram style clustering; novel lattice rescoring methods; recurrent neural network language models; speech recognition systems; strong generalization performance; vector representation; Artificial neural networks; Context; Decoding; History; Lattices; Recurrent neural networks; Vectors; language model; recurrent neural network; speech recognition;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2014 IEEE International Conference on
Conference_Location :
Florence
DOI :
10.1109/ICASSP.2014.6854535