Title of article :
Probabilistic classifiers with a generalized Gaussian scale mixture prior
Author/Authors :
Liu، نويسنده , , Guoqing and Wu، نويسنده , , Jianxin and Zhou، نويسنده , , Suiping، نويسنده ,
Issue Information :
روزنامه با شماره پیاپی سال 2013
Pages :
14
From page :
332
To page :
345
Abstract :
Most of the existing probabilistic classifiers are based on sparsity-inducing modeling. However, we show that sparsity is not always desirable in practice, and only an appropriate degree of sparsity is profitable. In this work, we propose a flexible probabilistic model using a generalized Gaussian scale mixture (GGSM) prior that can provide an appropriate degree of sparsity for its model parameters, and yield either sparse or non-sparse estimates according to the intrinsic sparsity of features in a dataset. Model learning is carried out by an efficient modified maximum a posterior (MAP) estimate. We also show relationships of the proposed model to existing probabilistic classifiers as well as iteratively re-weighted l1 and l2 minimizations. We then study different types of likelihood working with the GGSM prior in kernel-based setup, based on which an improved kernel-based GGIG is presented. Experiments demonstrate that the proposed method has better or comparable performances in linear classifiers as well as in kernel-based classification.
Keywords :
Classification , Prior distribution , Likelihood function , Generalized Gaussian scale mixture
Journal title :
PATTERN RECOGNITION
Serial Year :
2013
Journal title :
PATTERN RECOGNITION
Record number :
1735105
Link To Document :
بازگشت