DocumentCode :
1687616
Title :
Kernelized log linear models for continuous speech recognition
Author :
Shi-Xiong Zhang ; Gales, Mark J.F.
Author_Institution :
Dept. of Eng., Univ. of Cambridge, Cambridge, UK
fYear :
2013
Firstpage :
6950
Lastpage :
6954
Abstract :
Large margin criteria and discriminative models are two effective improvements for HMM-based speech recognition. This paper proposed a large margin trained log linear model with kernels for CSR. To avoid explicitly computing in the high dimensional feature space and to achieve the nonlinear decision boundaries, a kernel based training and decoding framework is proposed in this work. To make the system robust to noise a kernel adaptation scheme is also presented. Previous work in this area is extended in two directions. First, most kernels for CSR focus on measuring the similarity between two observation sequences. The proposed joint kernels defined a similarity between two observation-label sequence pairs on the sentence level. Second, this paper addresses how to efficiently employ kernels in large margin training and decoding with lattices. To the best of our knowledge, this is the first attempt at using large margin kernel-based log linear models for CSR. The model is evaluated on a noise corrupted continuous digit task: AURORA 2.0.
Keywords :
hidden Markov models; speech recognition; CSR; HMM based speech recognition; Kernelized log linear models; continuous speech recognition; decoding framework; kernel based training; nonlinear decision boundaries; Computational modeling; Decoding; Hidden Markov models; Joints; Kernel; Speech recognition; Training; kernel; large margin; log linear model;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on
Conference_Location :
Vancouver, BC
ISSN :
1520-6149
Type :
conf
DOI :
10.1109/ICASSP.2013.6639009
Filename :
6639009
Link To Document :
بازگشت