Title :
The shared dirichlet priors for bayesian language modeling
Author_Institution :
Dept. of Electr. & Comput. Eng., Nat. Chiao Tung Univ., Hsinchu, Taiwan
Abstract :
We present a new full Bayesian approach for language modeling based on the shared Dirichlet priors. This model is constructed by introducing the Dirichlet distribution to represent the uncertainty of n-gram parameters in training phase as well as in test time. Given a set of training data, the marginal likelihood over n-gram probabilities is illustrated in a form of linearly-interpolated n-grams. The hyperparameters in Dirichlet distributions are interpreted as the prior backoff information which is shared for the group of n-gram histories. This study estimates the shared hyperparameters by maximizing the marginal distribution of n-gram given the training data. Such Bayesian language model is connected to the smoothed language model. Experimental results show the superiority of the proposed method to the other methods in terms of perplexity and word error rate.
Keywords :
Bayes methods; interpolation; natural language processing; speech processing; Bayesian language model; Bayesian language modeling; Dirichlet distributions; backoff information; marginal likelihood; n-gram histories; n-gram parameters; n-gram probabilities; natural language; perplexity error rate; shared Dirichlet priors; smoothed language model; speech processing; training data; word error rate; Artificial neural networks; Bayes methods; Computational modeling; History; Interpolation; Smoothing methods; Training data; Bayesian learning; language model; model smoothing; optimal hyperparameter;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location :
South Brisbane, QLD
DOI :
10.1109/ICASSP.2015.7178337