Projected Langevin Monte Carlo algorithms in non-convex and super-linear setting
Abstract: It is of significant interest in many applications to sample from a high-dimensional target distribution $\pi$ with the density $\pi(\text{d} x) \propto e{-U(x)} (\text{d} x) $, based on the temporal discretization of the Langevin stochastic differential equations (SDEs). In this paper, we propose an explicit projected Langevin Monte Carlo (PLMC) algorithm with non-convex potential $U$ and super-linear gradient of $U$ and investigate the non-asymptotic analysis of its sampling error in total variation distance. Equipped with time-independent regularity estimates for the associated Kolmogorov equation, we derive the non-asymptotic bounds on the total variation distance between the target distribution of the Langevin SDEs and the law induced by the PLMC scheme with order $\mathcal{O}(d{\max{3\gamma/2 , 2\gamma-1 }} h |\ln h|)$, where $d$ is the dimension of the target distribution and $\gamma \geq 1$ characterizes the growth of the gradient of $U$. In addition, if the gradient of $U$ is globally Lipschitz continuous, an improved convergence order of $\mathcal{O}(d{3/2} h)$ for the classical Langevin Monte Carlo (LMC) scheme is derived with a refinement of the proof based on Malliavin calculus techniques. To achieve a given precision $\epsilon$, the smallest number of iterations of the PLMC algorithm is proved to be of order ${\mathcal{O}}\big(\tfrac{d{\max{3\gamma/2 , 2\gamma-1 }}}{\epsilon} \ \cdot \ln (\tfrac{d}{\epsilon}) \cdot \ln (\tfrac{1}{\epsilon}) \big)$. In particular, the classical Langevin Monte Carlo (LMC) scheme with the non-convex potential $U$ and the globally Lipschitz gradient of $U$ can be guaranteed by order ${\mathcal{O}}\big(\tfrac{d{3/2}}{\epsilon} \cdot \ln (\tfrac{1}{\epsilon}) \big)$. Numerical experiments are provided to confirm the theoretical findings.
- Stochastic C-stability and B-consistency of explicit and implicit Euler-type schemes. Journal of Scientific Computing, 67(3):955–987, 2016.
- Stochastic C-stability and B-consistency of explicit and implicit Milstein-type schemes. Journal of Scientific Computing, 70(3):1042–1077, 2017.
- C.-E. Bréhier. Approximation of the invariant measure with an Euler scheme for stochastic PDEs driven by space-time white noise. Potential Analysis, 40(1):1–40, 2014.
- The tamed unadjusted Langevin algorithm. Stochastic Processes and their Applications, 129(10):3638–3663, 2019.
- S. Cerrai. Second order PDE’s in finite and infinite dimension: a probabilistic approach. Springer, 2001.
- Sharp convergence rates for Langevin dynamics in the nonconvex setting. arXiv preprint arXiv:1805.01648, 2018.
- G. Da Prato. An introduction to infinite-dimensional analysis. Springer Science & Business Media, 2006.
- A. S. Dalalyan. Theoretical guarantees for approximate sampling from smooth and log-concave densities. Journal of the Royal Statistical Society Series B: Statistical Methodology, 79(3):651–676, 2017.
- A. Durmus and É. Moulines. Nonasymptotic convergence analysis for the unadjusted Langevin algorithm. The Annals of Applied Probability, 27(3):1551–1587, 2017.
- A. Durmus and É. Moulines. High-dimensional Bayesian inference via the unadjusted Langevin algorithm. Bernoulli, 25(4A):2854–2882, 2019.
- Formulae for the derivatives of heat semigroups. Journal of Functional Analysis, 125(1):252–286, 1994.
- B. Goldys. Exponential Ergodicity for Stochastic Reaction–Diffusion Equations. Stochastic Partial Differential Equations and Applications-VII, page 115, 2005.
- Strong and weak divergence in finite time of Euler’s method for stochastic differential equations with non-globally Lipschitz continuous coefficients. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 467(2130):1563–1576, 2011.
- Strong convergence of an explicit numerical method for SDEs with nonglobally Lipschitz continuous coefficients. The Annals of Applied Probability, 22(4):1611–1641, 2012.
- Sqrt (d) Dimension Dependence of Langevin Monte Carlo. In The International Conference on Learning Representations, 2022.
- Explicit numerical approximations for stochastic differential equations in finite and infinite horizons: truncation methods, convergence in pth moment and stability. IMA Journal of Numerical Analysis, 39(2):847–892, 2019.
- Unadjusted Langevin Algorithms for SDEs with Hölder Drift. arXiv preprint arXiv:2310.00232, 2023.
- Nonasymptotic bounds for sampling algorithms without log-concavity. The Annals of Applied Probability, 30(4):1534–1581, 2020.
- Ergodicity for SDEs and approximations: locally Lipschitz vector fields and degenerate noise. Stochastic processes and their applications, 101(2):185–232, 2002.
- Non-asymptotic convergence bounds for modified tamed unadjusted Langevin algorithm in non-convex setting. arXiv preprint arXiv:2207.02600, 2022.
- G. Pages and F. Panloup. Unadjusted Langevin algorithm with multiplicative noise: Total variation and Wasserstein bounds. The Annals of Applied Probability, 33(1):726–779, 2023.
- Linear implicit approximations of invariant measures of semi-linear SDEs with non-globally Lipschitz coefficients. arXiv preprint arXiv:2308.12886, 2023.
- S. Sabanis. Euler approximations with varying coefficients: the case of superlinearly growing diffusion coefficients. The Annals of Applied Probability, 26(4):2083–2105, 2016.
- Ł. Szpruch and X. Zhāng. V-integrability, asymptotic stability and comparison property of explicit numerical schemes for non-linear SDEs. Mathematics of Computation, 87(310):755–783, 2018.
- X. Wang and S. Gan. The tamed Milstein method for commutative stochastic differential equations with non-globally Lipschitz continuous coefficients. Journal of Difference Equations and Applications, 19(3):466–490, 2013.
- Mean-square convergence rates of stochastic theta methods for SDEs under a coupled monotonicity condition. BIT Numerical Mathematics, 60(3):759–790, 2020.
- Weak error analysis for strong approximation schemes of SDEs with super-linear coefficients. IMA Journal of Numerical Analysis, page drad083, 11 2023.
- Second-order numerical methods of weak convergence for SDEs with super-linear coefficients. preprint, 2023.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.