August 2021 Is there an analog of Nesterov acceleration for gradient-based MCMC?
Yi-An Ma, Niladri S. Chatterji, Xiang Cheng, Nicolas Flammarion, Peter L. Bartlett, Michael I. Jordan
Author Affiliations +
Bernoulli 27(3): 1942-1992 (August 2021). DOI: 10.3150/20-BEJ1297

Abstract

We formulate gradient-based Markov chain Monte Carlo (MCMC) sampling as optimization on the space of probability measures, with Kullback–Leibler (KL) divergence as the objective functional. We show that an underdamped form of the Langevin algorithm performs accelerated gradient descent in this metric. To characterize the convergence of the algorithm, we construct a Lyapunov functional and exploit hypocoercivity of the underdamped Langevin algorithm. As an application, we show that accelerated rates can be obtained for a class of nonconvex functions with the Langevin algorithm.

Citation

Download Citation

Yi-An Ma. Niladri S. Chatterji. Xiang Cheng. Nicolas Flammarion. Peter L. Bartlett. Michael I. Jordan. "Is there an analog of Nesterov acceleration for gradient-based MCMC?." Bernoulli 27 (3) 1942 - 1992, August 2021. https://doi.org/10.3150/20-BEJ1297

Information

Received: 1 October 2019; Revised: 1 October 2020; Published: August 2021
First available in Project Euclid: 10 May 2021

Digital Object Identifier: 10.3150/20-BEJ1297

Keywords: accelerated gradient descent , Langevin Monte Carlo , Markov chain Monte Carlo , sampling algorithms

Rights: Copyright © 2021 ISI/BS

JOURNAL ARTICLE
51 PAGES

This article is only available to subscribers.
It is not available for individual sale.
+ SAVE TO MY LIBRARY

Vol.27 • No. 3 • August 2021
Back to Top