Markov Chain Monte Carlo for Computer Vision
--- A tutorial at ICCV05 by Zhu, Delleart and Tu
Markov chain Monte Carlo is a general computing technique that has been widely used in physics, chemistry, biology, statistics, and computer science. It simulates a Markov chain whose invariant states follow a given (target) probability in a very high (say millions) dimensional state space. Essentially, it generates fair samples from a probability which are used for many purposes.A. System simulation. For many systems, their states are thought to follow some probability models. e.g. in statistical physics, the microscopic states of a system follows a Gibbs model given the macroscopic constraints. Sometimes, we have hard constraints and a uniform probability on the set of valid states. The fair samples produced by MCMC will show us what states are typical of the underlying system. E.g. the typical protein folding under certain conditions. In computer vision, this is often called "synthesis" ---the visual apparance of the simulated images, textures, and shapes is a way to verify the sufficiency of the underlying model. So it also provides a tool for learning and model verification.B. Scientific computing. In scientific computing, one often needs to compute the integral in very high dimensional space. E.g. the expected (macroscopic) property of a system. This is often done by Monte Carlo integration, i.e. estimating the expectation by empirical (sample) mean. Another interesting problem is approximate counting. E.g. how many non-self-intersecting paths are in a 2D n x n lattice? One may also estimate the value of pi if we can generate uniform samples in a unit square. In computer vision, Monte Carlo integration is used in learning and model estimation, and has also been used in motion tracking etc. C. Optimization and Bayesian inference. The objective is to compute the global optimum of some Bayesian posterior probability. In vision, a Bayesian posterior is often quite "peaky" or "very cold" if it is well formulated. This is why human vision has only one or a few of interpretations (otherwise we are very confused). Thus simulating the posterior will yield the plausible visual interpretation of the image. Very often the MCMC needs to go with an simulated annealing procedure. In the following, we prepare 7 lectures that are most related to vision research. This is intended for a tutorial by Zhu, Tu and Delleart at ICCV05.
1. Introduction to Markov Chain Monte Carlo (1) What is Markov chain Monte Carlo? (2) Why using MCMC? (3) Examples (4) Computing with two categories of models in vision: (5) Brief history of MCMC 2. MCMC Basics--Metropolis, Metropolis-Hastings, and Gibbs Sampling (1) Markov Chains (2) Inference and Estimation via Sampling (3) The Gibbs Sampler (4) Metropolis-Hastings (5) Metropolis and Gibbs (revisited) 3. A variety of tricks for MCMC design (1) Hit-and-run, random ray (2) Generalized Gibbs sampler (3) Simulated tempering (4) Data augmentation (5) Slice sampling (6) Cluster sampling (7) Metropolized Gibbs Sampler 4. Convergence and Exact sampling techniques (1) Definitions and terminologies. (2) Exact sampling techniques (3) Convergence rate and bounds using eigen-based analysis. (4) First hitting time analysis: ways to analyze the proposals. 5. Trans-dimensional MCMC (1) Model Selection (2) Reversible Jump MCMC (3) Regression Example (4) Rao-Blackwellized Sampling (5) Polygonal Random Fields 6. Cluster sampling with SW-cut (1) Graph partition and labeling (2) Clustering with bottom-up edge probabilities (3) Moving in the partition space (4) Swendsen-Wang Cuts (5) Examples 7. Data-driven MCMC: Integrating MCMC Search with Discriminative Computing (1) Generative and discriminative models (2) DDMCMC case study (3) DDMCMC approaches for vision applications
Some reference books: [1]. J.S. Liu, Monte Carlo Strategies in Scientific Computing, Springer, 2001. [2]. W. R. Gilks et al (eds) Markov Chain Monte Carlo in Practice, Chapman & Hall, 1996. [3]. P. Bremaud, Markov Chains: Gibbs Fields, Monte Carlo Simulation and Queues, Springer, 1999