Title :
Adapting n-gram maximum entropy language models with conditional entropy regularization
Author :
Rastrow, Ariya ; Dredze, Mark ; Khudanpur, Sanjeev
Author_Institution :
Human Language Technol. Center of Excellence, Johns Hopkins Univ., Baltimore, MD, USA
Abstract :
Accurate estimates of language model parameters are critical for building quality text generation systems, such as automatic speech recognition. However, text training data for a domain of interest is often unavailable. Instead, we use semi-supervised model adaptation; parameters are estimated using both unlabeled in-domain data (raw speech audio) and labeled out of domain data (text.) In this work, we present a new semi-supervised language model adaptation procedure for Maximum Entropy models with n-gram features. We augment the conventional maximum likelihood training criterion on out-of-domain text data with an additional term to minimize conditional entropy on in-domain audio. Additionally, we demonstrate how to compute conditional entropy efficiently on speech lattices using first- and second-order expectation semirings. We demonstrate improvements in terms of word error rate over other adaptation techniques when adapting a maximum entropy language model from broadcast news to MIT lectures.
Keywords :
entropy; maximum likelihood estimation; speech recognition; MIT lectures; automatic speech recognition; broadcast news; building quality text generation systems; conditional entropy regularization; conventional maximum likelihood training criterion; n-gram maximum entropy language models; semisupervised model adaptation;; unlabeled in-domain data; Adaptation models; Computational modeling; Data models; Entropy; Lattices; Speech; Training;
Conference_Titel :
Automatic Speech Recognition and Understanding (ASRU), 2011 IEEE Workshop on
Conference_Location :
Waikoloa, HI
Print_ISBN :
978-1-4673-0365-1
Electronic_ISBN :
978-1-4673-0366-8
DOI :
10.1109/ASRU.2011.6163934