DocumentCode :
2979366
Title :
A whole sentence maximum entropy language model
Author :
Rosenfeld, R.
Author_Institution :
Sch. of Comput. Sci., Carnegie Mellon Univ., Pittsburgh, PA, USA
fYear :
1997
fDate :
14-17 Dec 1997
Firstpage :
230
Lastpage :
237
Abstract :
Introduces a new kind of language model, which models whole sentences or utterances directly using the maximum entropy (ME) paradigm. The new model is conceptually simpler, and more naturally suited to modeling whole-sentence phenomena, than the conditional ME models proposed to date. By avoiding the chain rule, the model treats each sentence or utterance as a “bag of features”, where features are arbitrary computable properties of the sentence. The model is unnormalizable, but this does not interfere with training (done via sampling) or with use. Using the model is computationally straightforward. The main computational cost of training the model is in generating sample sentences from a Gibbs distribution. Interestingly, this cost has different dependencies, and is potentially lower than in the comparable conditional ME model
Keywords :
maximum entropy methods; natural languages; probability; Gibbs distribution; arbitrary computable properties; bag of features; chain rule; computational cost; sample sentence generation; sampling; training; unnormalizable model; utterances; whole-sentence maximum entropy language model; Computational Intelligence Society; Computational efficiency; Computational modeling; Computer science; Costs; Entropy; Exponential distribution; Probability; Sampling methods; Solid modeling;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Automatic Speech Recognition and Understanding, 1997. Proceedings., 1997 IEEE Workshop on
Conference_Location :
Santa Barbara, CA
Print_ISBN :
0-7803-3698-4
Type :
conf
DOI :
10.1109/ASRU.1997.659010
Filename :
659010
Link To Document :
بازگشت