Title :
Variable n-grams and extensions for conversational speech language modeling
Author :
Siu, Manhung ; Ostendorf, Mari
Author_Institution :
Hong Kong Univ. of Sci. & Technol., Kowloon, Hong Kong
fDate :
1/1/2000 12:00:00 AM
Abstract :
Recent progress in variable n-gram language modeling provides an efficient representation of n-gram models and makes training of higher order n grams possible. We apply the variable n-gram design algorithm to conversational speech, extending the algorithm to learn skips and context-dependent classes to handle conversational speech characteristics such as filler words, repetitions, and other disfluencies. Experiments show that using the extended variable n-gram results in a language model that captures 4-gram context with less than half the parameters of a standard trigram while also improving the test perplexity and recognition accuracy
Keywords :
computational linguistics; natural languages; speech recognition; context-dependent classes; conversational speech language modeling; experiments; speech recognition; training; trigram; variable n-gram language modeling; Algorithm design and analysis; Context modeling; Costs; Helium; History; Natural languages; Parameter estimation; Speech recognition; Testing; Vocabulary;
Journal_Title :
Speech and Audio Processing, IEEE Transactions on