DocumentCode
3432845
Title
Bayesian supervised learning with non-Gaussian latent variables
Author
Siwei Lyu
Author_Institution
Comput. Sci. Dept., Univ. at Albany, Albany, NY, USA
fYear
2013
fDate
6-10 July 2013
Firstpage
659
Lastpage
663
Abstract
We describe a Bayesian learning scheme for the hierarchal Bayesian linear model, which is based on the Gaussian scale mixture (GSM) modeling of the distribution of the latent variable. The proposed method takes advantage of the hierarchal Gaussian structure for a simple Monte-Carlo sampling algorithm. Particularly, with a single hidden scale parameter controlling the distribution of the latent variables, it leads to an efficient algorithm without explicit matrix inversion.
Keywords
Gaussian processes; Monte Carlo methods; belief networks; learning (artificial intelligence); sampling methods; Bayesian supervised learning; GSM modeling; Gaussian scale mixture modeling; Monte-Carlo sampling algorithm; hierarchal Bayesian linear model; hierarchal Gaussian structure; latent variable distribution; nonGaussian latent variables; Bayes methods; Computational modeling; Eigenvalues and eigenfunctions; Estimation; GSM; Monte Carlo methods; Vectors; Bayesian learning; Gaussian scale mixtures; latent variable models;
fLanguage
English
Publisher
ieee
Conference_Titel
Signal and Information Processing (ChinaSIP), 2013 IEEE China Summit & International Conference on
Conference_Location
Beijing
Type
conf
DOI
10.1109/ChinaSIP.2013.6625424
Filename
6625424
Link To Document