Accelerating Hamiltonian Monte Carlo via Chebyshev Integration Time

07/05/2022
by   Jun-Kun Wang, et al.
0

Hamiltonian Monte Carlo (HMC) is a popular method in sampling. While there are quite a few works of studying this method on various aspects, an interesting question is how to choose its integration time to achieve acceleration. In this work, we consider accelerating the process of sampling from a distribution π(x) ∝exp(-f(x)) via HMC via time-varying integration time. When the potential f is L-smooth and m-strongly convex, i.e. for sampling from a log-smooth and strongly log-concave target distribution π, it is known that under a constant integration time, the number of iterations that ideal HMC takes to get an ϵ Wasserstein-2 distance to the target π is O( κlog1/ϵ ), where κ := L/m is the condition number. We propose a scheme of time-varying integration time based on the roots of Chebyshev polynomials. We show that in the case of quadratic potential f, i.e., when the target π is a Gaussian distribution, ideal HMC with this choice of integration time only takes O( √(κ)log1/ϵ ) number of iterations to reach Wasserstein-2 distance less than ϵ; this improvement on the dependence on condition number is akin to acceleration in optimization. The design and analysis of HMC with the proposed integration time is built on the tools of Chebyshev polynomials. Experiments find the advantage of adopting our scheme of time-varying integration time even for sampling from distributions with smooth strongly convex potentials that are not quadratic.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset