DocumentCode :
3444
Title :
Vector quantization of LSF parameters with a mixture of dirichlet distributions
Author :
Zhanyu Ma ; Leijon, Arne ; Kleijn, W. Bastiaan
Author_Institution :
Pattern Recognition & Intell. Syst. Lab., Beijing Univ. of Posts & Telecommun., Beijing, China
Volume :
21
Issue :
9
fYear :
2013
fDate :
Sept. 2013
Firstpage :
1777
Lastpage :
1790
Abstract :
Quantization of the linear predictive coding parameters is an important part in speech coding. Probability density function (PDF)-optimized vector quantization (VQ) has been previously shown to be more efficient than VQ based only on training data. For data with bounded support, some well-defined bounded-support distributions (e.g., the Dirichlet distribution) have been proven to outperform the conventional Gaussian mixture model (GMM), with the same number of free parameters required to describe the model. When exploiting both the boundary and the order properties of the line spectral frequency (LSF) parameters, the distribution of LSF differences LSF can be modelled with a Dirichlet mixture model (DMM). We propose a corresponding DMM based VQ. The elements in a Dirichlet vector variable are highly mutually correlated. Motivated by the Dirichlet vector variable´s neutrality property, a practical non-linear transformation scheme for the Dirichlet vector variable can be obtained. Similar to the Karhunen-Loève transform for Gaussian variables, this non-linear transformation decomposes the Dirichlet vector variable into a set of independent beta-distributed variables. Using high rate quantization theory and by the entropy constraint, the optimal inter- and intra-component bit allocation strategies are proposed. In the implementation of scalar quantizers, we use the constrained-resolution coding to approximate the derived constrained-entropy coding. A practical coding scheme for DVQ is designed for the purpose of reducing the quantization error accumulation. The theoretical and practical quantization performance of DVQ is evaluated. Compared to the state-of-the-art GMM-based VQ and recently proposed beta mixture model (BMM) based VQ, DVQ performs better, with even fewer free parameters and lower computational cost
Keywords :
Gaussian processes; Karhunen-Loeve transforms; entropy codes; linear predictive coding; speech coding; vector quantisation; BMM; DMM based VQ; Dirichlet distributions; Dirichlet mixture model; Dirichlet vector variable; GMM; Gaussian mixture model; Gaussian variables; Karhunen-Loève transform; LSF parameters; PDF; beta mixture model; bounded-support distributions; constrained-entropy coding; constrained-resolution coding; entropy constraint; high rate quantization theory; independent beta-distributed variables; line spectral frequency; linear predictive coding parameters; nonlinear transformation; optimal intercomponent bit allocation strategies; optimal intracomponent bit allocation strategies; probability density function; quantization error accumulation reduction; scalar quantizers; speech coding; vector quantization; Bit rate; Computational efficiency; Computational modeling; Encoding; Quantization; Training data; Vectors; Beta distribution; Dirichlet distribution; Speech coding; bounded support distribution; line spectral frequency; mixture modelling; neutrality property; vector quantization;
fLanguage :
English
Journal_Title :
Audio, Speech, and Language Processing, IEEE Transactions on
Publisher :
ieee
ISSN :
1558-7916
Type :
jour
DOI :
10.1109/TASL.2013.2238732
Filename :
6407842
Link To Document :
بازگشت