Title :
A whole sentence maximum entropy language model
Author_Institution :
Sch. of Comput. Sci., Carnegie Mellon Univ., Pittsburgh, PA, USA
Abstract :
Introduces a new kind of language model, which models whole sentences or utterances directly using the maximum entropy (ME) paradigm. The new model is conceptually simpler, and more naturally suited to modeling whole-sentence phenomena, than the conditional ME models proposed to date. By avoiding the chain rule, the model treats each sentence or utterance as a “bag of features”, where features are arbitrary computable properties of the sentence. The model is unnormalizable, but this does not interfere with training (done via sampling) or with use. Using the model is computationally straightforward. The main computational cost of training the model is in generating sample sentences from a Gibbs distribution. Interestingly, this cost has different dependencies, and is potentially lower than in the comparable conditional ME model
Keywords :
maximum entropy methods; natural languages; probability; Gibbs distribution; arbitrary computable properties; bag of features; chain rule; computational cost; sample sentence generation; sampling; training; unnormalizable model; utterances; whole-sentence maximum entropy language model; Computational Intelligence Society; Computational efficiency; Computational modeling; Computer science; Costs; Entropy; Exponential distribution; Probability; Sampling methods; Solid modeling;
Conference_Titel :
Automatic Speech Recognition and Understanding, 1997. Proceedings., 1997 IEEE Workshop on
Conference_Location :
Santa Barbara, CA
Print_ISBN :
0-7803-3698-4
DOI :
10.1109/ASRU.1997.659010