|Seminar on Theoretical Machine Learning|
|Topic:||Beyond log-concavity: provable guarantees for sampling multi-modal distributions using simulated tempering Langevin Monte Carlo|
|Date:||Monday, November 27|
|Time/Room:||12:30pm - 1:45pm/White-Levy Room|
A fundamental problem in Bayesian statistics is sampling from distributions that are only specified up to a partition function (constant of proportionality). In particular, we consider the problem of sampling from a distribution given access to the gradient of the log-pdf. For log-concave distributions, classical results due to Bakry and Emery show that natural continuous-time Markov chains called Langevin diffusions mix in polynomial time. But in practice, distributions are often multi-modal and hence non-log-concave, and can take exponential time to mix. We address this problem by combining Langevin diffusion with simulated tempering. The result is a Markov chain that mixes in polynomial rather than exponential time by transitioning between different temperatures of the distribution. We prove fast mixing for any distribution that is close to a mixture of gaussians of equal variance. Based on the paper https://arxiv.org/abs/1710.02736. Joint work with Rong Ge (Duke) and Andrej Risteski (MIT). To appear in NIPS AABI workshop.