Title :
Hierarchical linear discounting class N-gram language models: A multilevel class hierarchy approach
Author :
Zitouni, Imed ; Zhou, Qiru
Author_Institution :
TJ. Watson Res. Center, IBM, Yorktown Heights, NY
fDate :
March 31 2008-April 4 2008
Abstract :
We introduce in this paper a hierarchical linear discounting class n-gram language modeling technique that has the advantage of combining several language models, trained at different nodes in a class hierarchy. The approach hierarchically clusters the word vocabulary into a word-tree. The closer a tree node is to the leaves, the more specific the corresponding word class is. The tree is used to balance generalization ability and word specificity when estimating the likelihood of an n-gram event. Experiments are conducted on Wall Street Journal corpus using a vocabulary of 20,000 words. Results show a reduction on the test perplexity over the standard n-gram approaches by 10%. We also report considerable improvement in the accuracy of the speech recognition task.
Keywords :
speech recognition; hierarchical linear discounting class n-gram language models; multilevel class hierarchy approach; speech recognition task; word tree; Frequency; Interpolation; Smoothing methods; Speech recognition; Testing; Vocabulary; Class Hierarchy; Language Modeling; Linear Distortion; n-gram;
Conference_Titel :
Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on
Conference_Location :
Las Vegas, NV
Print_ISBN :
978-1-4244-1483-3
Electronic_ISBN :
1520-6149
DOI :
10.1109/ICASSP.2008.4518760