Issue No. 12 - December (2008 vol. 30)
Jaemo Sung , POSTECH, Pohang
Zoubin Ghahramani , University of Cambridge, Cambridge
Sung-Yang Bang , POSTECH, Pohang
Variational Bayesian Expectation-Maximization (VBEM), an approximate inference method for probabilistic models based on factorizing over latent variables and model parameters, has been a standard technique for practical Bayesian inference. In this paper, we introduce a more general approximate inference framework for conjugate-exponential family models, which we call Latent-Space Variational Bayes (LSVB). In this approach, we integrate out model parameters in an exact way, leaving only the latent variables. It can be shown that the LSVB approach gives better estimates of the model evidence as well as the distribution over latent variables than the VBEM approach, but in practice, the distribution over latent variables has to be approximated. As a practical implementation, we present a First-order LSVB (FoLSVB) algorithm to approximate this distribution over latent variables. From this approximate distribution, one can estimate the model evidence and the posterior over model parameters. The FoLSVB algorithm is directly comparable to the VBEM algorithm and has the same computational complexity. We discuss how LSVB generalizes the recently proposed collapsed variational methods  to general conjugate-exponential families. Examples based on mixtures of Gaussians and mixtures of Bernoullis with synthetic and real-world data sets are used to illustrate some advantages of our method over VBEM.
variational Bayesian inference, Machine learning, unsupervised Learning, latent variable model, conjugate exponential family, variational method, mixture of Gaussians
Z. Ghahramani, J. Sung and S. Bang, "Latent-Space Variational Bayes," in IEEE Transactions on Pattern Analysis & Machine Intelligence, vol. 30, no. , pp. 2236-2242, 2008.