DocumentCode :
1692042
Title :
Unifying PLDA and polynomial kernel SVMS
Author :
Yaman, Sibel ; Pelecanos, Jason ; Weizhong Zhu
Author_Institution :
IBM T. J. Watson Res. Center, Yorktown Heights, NY, USA
fYear :
2013
Firstpage :
7698
Lastpage :
7701
Abstract :
Probabilistic linear discriminant analysis (PLDA) is a generative model to explain between and within class variations. When the underlying latent variables are modelled by standard Gaussian distributions, the PLDA recognition scores can be evaluated as a dot product between a high dimensional PLDA feature vector and a weight vector. A key contribution of this paper is showing that the high dimensional PLDA feature vectors can be equivalently (in a non-strict sense) represented as the second-degree polynomial kernel induced features of the vectors formed by concatenating the two input vectors constituting a trial. This equivalence relationship paves the way for the speaker recognition problem to be viewed as a two-class support vector machine (SVM) training problem where higher degree polynomial kernels can give better discriminative power. To alleviate the large scale SVM training problem, we propose a kernel evaluation trick that greatly simplifies the kernel evaluation operations. In our experiments, a combination of multiple second degree polynomial kernel SVMs performed comparably to a state-of-the-art PLDA system. For the analysed test case, SVMs trained with third degree polynomial kernel reduced the EERs on average by 10% relative to that of the SVMs trained with second degree polynomial kernel.
Keywords :
Gaussian distribution; polynomials; probability; speaker recognition; support vector machines; EER; PLDA recognition scores; SVM training problem; class variations; discriminative power; dot product; equivalence relationship; generative model; high dimensional PLDA feature vector; higher degree polynomial kernels; kernel evaluation operations; latent variables; polynomial kernel SVMS; probabilistic linear discriminant analysis; second-degree polynomial kernel; speaker recognition; standard Gaussian distributions; third degree polynomial kernel; two-class support vector machine; weight vector; Kernel; Mathematical model; Polynomials; Speaker recognition; Support vector machines; Training; Vectors; PLDA; large scale SVMs; speaker recognition;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on
Conference_Location :
Vancouver, BC
ISSN :
1520-6149
Type :
conf
DOI :
10.1109/ICASSP.2013.6639161
Filename :
6639161
Link To Document :
بازگشت