Abstract
Hidden Markov random field models provide an appealing representation of images and other spatial problems. The drawback is that inference is not straightforward for these models as the normalisation constant for the likelihood is generally intractable except for very small observation sets. Variational methods are an emerging tool for Bayesian inference and they have already been successfully applied in other contexts. Focusing on the particular case of a hidden Potts model with Gaussian noise, we show how variational Bayesian methods can be applied to hidden Markov random field inference. To tackle the obstacle of the intractable normalising constant for the likelihood, we explore alternative estimation approaches for incorporation into the variational Bayes algorithm. We consider a pseudo-likelihood approach as well as the more recent reduced dependence approximation of the normalisation constant. To illustrate the effectiveness of these approaches we present empirical results from the analysis of simulated datasets. We also analyse a real dataset and compare results with those of previous analyses as well as those obtained from the recently developed auxiliary variable MCMC method and the recursive MCMC method. Our results show that the variational Bayesian analyses can be carried out much faster than the MCMC analyses and produce good estimates of model parameters. We also found that the reduced dependence approximation of the normalisation constant outperformed the pseudo-likelihood approximation in our analysis of real and synthetic datasets.
Article PDF
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
References
Attias, H.: Inferring parameters and structure of latent variable models by variational Bayes. In: Proceedings of 15th Conference on Uncertainty in Artificial Intelligence (1999)
Beal, M.J., Ghahramani, Z.: The variational Bayesian EM algorithm for incomplete data: with application to scoring graphical model structures. In: Bernardo, J.M., Bayarri, M.J., Berger, J.O., David, A.P., Heckerman, D., Smith, A.F.M., West, M. (eds.) Proceedings of the Seventh Valencia International Meeting. Bayesian Statistics, vol. 7, pp. 453–464. Oxford University Press, London (2003)
Besag, J.: Spatial interaction and the statistical analysis of lattice systems (with discussion). J. R. Stat. Soc. Ser. B 36, 192–236 (1974)
Besag, J.: Statistical analysis of non-lattice data. Statistician 24, 179–195 (1975)
Besag, J.: On the statistical analysis of dirty pictures (with discussion). J. R. Stat. Soc. Ser. B 48, 259–302 (1986)
Besag, J., York, J., Mollie, A.: Bayesian image restoration, with two applications in spatial statistics (with discussion). Ann. Inst. Stat. Math. 43, 1–59 (1991)
Besag, J., Higdon, D.: Bayesian analysis of agricultural field experiments. J. R. Stat. Soc. Ser. B 61, 691–746 (1999)
Buck, C.E., Cavanagh, W.G., Litton, C.D.: The spatial analysis of site phosphate data. In: Rhatz, S.P.Q. (ed.) Computer Applications and Quantitive Methods in Archeology. British Archaeological Reports, International Series, vol. 446. BAR, Oxford (1988)
Corduneanu, A., Bishop, C.M.: Variational Bayesian model selection for mixture distributions. In: Jaakkola, T., Richardson, T. (eds.) Artificial Intelligence and Statistics, pp. 27–34. Morgan Kaufmann, San Mateo (2001)
Friel, N., Pettitt, A.N., Reeves, R., Wit, E.: Bayesian inference in hidden Markov random fields for binary data defined on large lattices (2008, submitted)
Gelman, A., Meng, X.L.: Simulating normalizing constants: from importance sampling to bridge sampling to path sampling. Stat. Sci. 13, 163–185 (1998)
Geman, S., Geman, D.: Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Trans. Pattern Anal. Mach. Intell. 6, 721–741 (1984)
Gottardo, R., Besag, J., Stephens, M., Murua, A.: Probabilistic segmentation and intensity estimation for microarray images. Biostatistics 7, 85–89 (2006)
Green, P.J., Richardson, S.: Hidden Markov models and disease mapping. J. Am. Stat. Assoc. 97, 1055–1070 (2002)
Jordan, M.: Graphical models. Stat. Sci. 19, 140–155 (2004)
MacKay, D.J.C.: Ensemble learning for hidden Markov models. Technical Report, Cavendish Laboratory, University of Cambridge (1997)
McGrory, C.A.: Variational Approximations in Bayesian Model Selection. Ph.D. Thesis, University of Glasgow, UK (2005)
McGrory, C.A., Titterington, D.M.: Variational approximations in Bayesian model selection for finite mixture distributions. Comput. Stat. Data Anal. 51, 5352–5367 (2007)
McGrory, C.A., Titterington, D.M.: Bayesian analysis of hidden Markov models using variational approximations. Aust. N. Z. J. Stat. (2008, to appear)
Møller, J., Pettitt, A.N., Reeves, R., Berthelsen, K.K.: An efficient Markov chain Monte Carlo method for distributions with intractable normalising constants. Biometrika 93, 451–458 (2006)
Murray, I., Ghahramani, Z., MacKay, D.J.C.: MCMC for doubly-intractable distributions. In: Proceedings of 22nd Conference on Uncertainty in Artificial Intelligence (2006)
Newman, M.E.J., Barkema, G.T.: Monte Carlo Methods in Statistical Physics. Oxford University Press, London (1999)
Pettitt, A.N., Friel, N., Reeves, R.: Efficient calculation of the normalising constant of the autologistic and related models on the cylinder and lattice. J. R. Stat. Soc. Ser. B 65, 235–247 (2003)
Reeves, R., Pettitt, A.N.: Efficient recursions for general factorisable models. Biometrika 91, 751–757 (2004)
Rydén, T., Titterington, D.M.: Computational Bayesian analysis of hidden Markov models. J. Comput. Graph. Stat. 7, 194–211 (1998)
Smith, M., Fahrmeir, L.: Spatial Bayesian variable selection with application to functional magnetic resonance imaging. J. Am. Stat. Assoc. 102, 417–431 (2007)
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
McGrory, C.A., Titterington, D.M., Reeves, R. et al. Variational Bayes for estimating the parameters of a hidden Potts model. Stat Comput 19, 329–340 (2009). https://doi.org/10.1007/s11222-008-9095-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11222-008-9095-6