Title :
Optimal quantization and bit allocation for compressing large discriminative feature space transforms
Author :
Marcheret, Etienne ; Goel, Vaibhava ; Olsen, Peder A.
Author_Institution :
IBM T. J. Watson Res., Yorktown Heights, NY, USA
fDate :
Nov. 13 2009-Dec. 17 2009
Abstract :
Discriminative training of the feature space using the minimum phone error (MPE) objective function has been shown to yield remarkable accuracy improvements. These gains, however, come at a high cost of memory required to store the transform. In a previous paper we reduced this memory requirement by 94% by quantizing the transform parameters. We used dimension dependent quantization tables and learned the quantization values with a fixed assignment of transform parameters to quantization values. In this paper we refine and extend the techniques to attain a further 35% reduction in memory with no degradation in sentence error rate. We discuss a principled method to assign the transform parameters to quantization values. We also show how the memory can be gradually reduced using a Viterbi algorithm to optimally assign variable number of bits to dimension dependent quantization tables. The techniques described could also be applied to the quantization of general linear transforms - a problem that should be of wider interest.
Keywords :
speech recognition; Viterbi algorithm; bit allocation; discriminative feature space transforms; minimum phone error; optimal quantization; sentence error rate; Automatic speech recognition; Automotive engineering; Bit rate; Compaction; Costs; Degradation; Error analysis; Gaussian processes; Quantization; Viterbi algorithm;
Conference_Titel :
Automatic Speech Recognition & Understanding, 2009. ASRU 2009. IEEE Workshop on
Conference_Location :
Merano
Print_ISBN :
978-1-4244-5478-5
Electronic_ISBN :
978-1-4244-5479-2
DOI :
10.1109/ASRU.2009.5373407