Papers
Topics
Authors
Recent
Search
2000 character limit reached

Weak Convergence Rate Bounds

Updated 11 December 2025
  • Weak Convergence Rate Bounds are quantitative estimates that measure how fast the expectation of test functionals converges in discretized stochastic and numerical systems.
  • They are crucial in applications such as SDEs, SPDEs, inverse problems, and optimization, offering rigorous error control when strong convergence is unattainable.
  • These bounds leverage analytical tools like Kolmogorov equations, weak Poincaré inequalities, and spectral regularization to yield precise rate estimates in diverse settings.

Weak convergence rate bounds provide explicit quantitative estimates on how fast sequences of measures, stochastic processes, operator semigroups, or iterates of optimization algorithms converge in law or against a prescribed class of test functions, typically as a function of algorithmic parameters, discretization granularity, or small parameters in multiscale dynamics. These bounds control the decay of expectation errors for functionals of the solution—not just pointwise trajectories—making them central in numerical analysis, stochastic process theory, and optimization, particularly when strong (pathwise) convergence is either unobtainable or too stringent. Weak rates are foundational for rigorous error analysis in PDEs, SDEs, SPDEs, inverse problems, adaptive MCMC, and statistical physics.

1. Fundamental Notions and Settings

Weak convergence rate bounds quantify the rate at which the expectation of a test functional φ\varphi under an approximate solution XnX^n converges to its value under the exact solution XX, i.e.,

E[φ(Xn)]E[φ(X)]Cr(n)|\mathbb{E}[\varphi(X^n)] - \mathbb{E}[\varphi(X)]| \leq C\,r(n)

for all φ\varphi in a class T\mathcal{T}, with r(n)0r(n)\to0 as nn\to\infty. The choice of T\mathcal{T} (bounded, Lipschitz, CkC^k), and the metric (total variation, Wasserstein, bounded Lipschitz, or Hilbert space dual-pairing) defines the "weakness" of the convergence.

Key settings in which weak convergence rates have been rigorously characterized include:

2. Weak Convergence Rate Bounds for Stochastic Differential Equations

Weak error bounds for SDE discretizations arise from expansions of the local error in expectation via the Kolmogorov backward equation, Malliavin calculus, or duality-based techniques. For example:

  • In the log-Heston model, under smooth payoff ff and Feller index ν>2\nu > 2, the drift-implicit Milstein–Euler scheme achieves a weak rate

E[f(XT)]E[f(xN)]=O(Δ)| \mathbb{E}[f(X_T)] - \mathbb{E}[f(x_N)]| = O(\Delta)

where Δ\Delta is the maximal step size (Altmayer et al., 2016). The analysis uses Kolmogorov PDE techniques and a Malliavin calculus integration-by-parts to achieve leading term cancellation.

  • For rough volatility models, the discretization error for smooth payoffs φCb3\varphi \in C_b^3 admits the lower bound

E[φ(ST)]E[φ(STn)]=O(n2H)=O(Δ2H)| \mathbb{E}[\varphi(S_T)] - \mathbb{E}[\varphi(S_T^n)] | = O(n^{-2H}) = O(\Delta^{2H})

where HH is the Hurst index of the volatility process (Bayer et al., 2022). If the volatility functional is linear, sharper bounds O(n(H+1/2))O(n^{-(H+1/2)}) are attainable.

For SPDEs such as the stochastic Allen–Cahn equation with additive noise, explicit full discretization (spectral Galerkin in space, tamed exponential Euler in time) yields weak convergence in both time and space with rate essentially double that of strong error: E[Φ(X(T))]E[Φ(XTM,N)]C(λN(γϵ)+τγϵ)|\mathbb{E}[\Phi(X(T))] - \mathbb{E}[\Phi(X_T^{M,N})]| \le C\left(\lambda_N^{-(\gamma-\epsilon)} + \tau^{\gamma-\epsilon}\right) for γ\gamma the noise regularity index and λN\lambda_N the NN-th Laplacian eigenvalue (Cai et al., 2019).

3. Weak Convergence via Functional Inequalities and Operator Semigroups

Degenerate or non-uniformly elliptic diffusions often lack classical Poincaré inequalities, precluding exponential convergence in strong metrics. Instead, weak Poincaré inequalities yield explicit subgeometric rate bounds:

  • For a C0C_0-semigroup PtP_t generated by L=SAL=S-A (symmetric SS, anti-symmetric AA), if weak Poincaré inequalities hold

Varπ(f)α(r)E(f,f)+rfosc2\mathrm{Var}_\pi(f) \le \alpha(r) \mathcal{E}(f,f) + r \|f\|_{\mathrm{osc}}^2

(with Dirichlet form E\mathcal{E}), then

supfosc1Varπ(Ptf)Ψ(t)\sup_{\|f\|_{\mathrm{osc}}\le1} \mathrm{Var}_\pi(P_t f) \leq \Psi(t)

where Ψ(t)\Psi(t) is given by an explicit minimization involving α1,α2\alpha_1,\alpha_2 and their rate of divergence as r0r\to0 (Grothaus et al., 2017, Bertram et al., 2021).

  • For polynomially confining potentials V1,V2V_1,V_2, decay can be stretched-exponential:

Varπ(Ptf)C1exp(C2tδ2/(4δ2+2d2+4))fosc2\mathrm{Var}_\pi(P_t f) \leq C_1 \exp\left(-C_2 t^{\delta_2/(4\delta_2+2d_2+4)}\right) \|f\|_{\mathrm{osc}}^2

and with weakly confining (logarithmic) potentials, only polynomial or iterated-logarithmic rates are obtained. The regime is dictated by the tails of the invariant measure (Bertram et al., 2021).

4. Weak Convergence in Inverse Problems and Spectral Regularization

In inverse problems, weak convergence rates quantify the decay of the bias and variance of linear observation functionals under spectral regularization. Key insights include:

  • For compact forward operators A:H1H2A:H_1\rightarrow H_2 and noisy samples at nn points, spectral regularization (e.g., Tikhonov, cutoff) admits weak error bounds for test functions ψD((A))\psi\in D((A^*)^{\dagger}):

Eψ,fλδ,nf=O(hd/2(hτd/2+λ1/2)+ν/(nλ))\mathbb{E}|\langle \psi, f^{\delta,n}_\lambda - f^\dagger \rangle| = O(h^{d/2}(h^{\tau-d/2} + \lambda^{1/2}) + \nu/(\sqrt{n}\lambda))

with h=O(n1/d)h = O(n^{-1/d}) the fill distance, τ\tau kernel smoothness, and ν\nu noise level. In trace-class settings, variance scaling improves and optimal weak rates are O(nmin(τ/d,2/3))O(n^{-\min(\tau/d,2/3)}) (Guastavino et al., 4 Dec 2025).

Sampling inequalities replace classical source conditions; thus, weak rates can be achieved independently of function smoothness assumptions, controlled solely by fill-distance/geometric considerations.

5. Weak Convergence in Stochastic Systems and Limit Theorems

Weak convergence rates also characterize fluctuation limits in multiscale stochastic systems:

  • In two-scale McKean–Vlasov models, after rescaling the deviation process,

Utϵ=XtϵXˉtϵU^\epsilon_t = \frac{X^\epsilon_t - \bar X_t}{\sqrt{\epsilon}}

it is shown that for smooth test functions ψ\psi,

suptTψ(Law(Utϵ))ψ(Law(Ut))CTϵ\sup_{t\le T} |\psi(\mathrm{Law}(U^\epsilon_t)) - \psi(\mathrm{Law}(U_t))| \leq C_T \epsilon

where UU is the unique limit McKean–Vlasov SDE (Xiang et al., 20 May 2025). This employs Poisson equation techniques and a Kolmogorov–Cauchy duality.

In directed polymer models, the weak-disorder regime (λ2(β)<ln(1/Td)\lambda_2(\beta)<\ln(1/T_d)) leads to polynomial L2L^2 rates for the partition function martingale: WWn2σ1(β)n(d2)/4,(n)\|W - W_n\|_2 \sim \sigma_1(\beta) n^{-(d-2)/4}, \quad (n\to\infty) with fluctuation scaling governed by a CLT with stable and mixing laws, in sharp contrast to the exponential rates seen on trees (Comets et al., 2016).

6. Weak Convergence Rates in Optimization and Sampling

In optimization and MCMC theory, "weak" convergence rates often pertain to Bregman divergences, Lyapunov function decay, or distances in Wasserstein/total variation metrics:

  • For mirror descent with μf\mu_f-strongly convex and LfL_f-smooth objective ff, and ϕ\phi likewise smooth and strongly convex, the Bregman divergence decays exponentially in continuous time:

Dϕ(x(t),x)Dϕ(x(0),x)eρt,ρ=ημf/LϕD_\phi(x(t),x^*) \le D_\phi(x(0),x^*) e^{-\rho t},\quad \rho = \eta\mu_f/L_\phi

and linearly in discrete time for optimal step size choices (Li et al., 2023).

  • More generally, Grimmer's meta-theorems show that general (Hölder) growth bounds allow for a systematic derivation of weak convergence rates in first-order methods and enable new lower bounds in rates under additional growth conditions. This unifies and extends the landscape of possible rates in convex optimization (Grimmer, 2021).

For adaptive MCMC, subgeometric weak convergence lower and upper bounds are matched up to log factors when the adaptation schedule G(t)G(t) decays fast: infQWd1(AQ(t),π)1/[Hw0,ϕ1(t)]κ/(ακ)\inf_{Q} W_{d\wedge1}(A_Q^{(t)},\pi) \gtrsim 1 / [H_{w_0,\phi}^{-1}(t)]^{\kappa/(\alpha-\kappa)} where the rate is determined by the drift function ϕ\phi and invariant measure tail κ\kappa; upper bounds of the same order obtain under minorization and drift conditions (Brown et al., 2024).

In data-driven generative modeling, weak log-concavity suffices for 2-Wasserstein non-asymptotic error bounds for probability flow ODE samplers: W2(Law(Z^T),p0)E0(init)+E1(discret.)+E2(score error)W_2(\mathrm{Law}(\widehat Z_T), p_0) \leq E_0(\text{init}) + E_1(\text{discret.}) + E_2(\text{score error}) with each term controlled in explicit dependence on step size hh, horizon TT, and other model parameters (Kremling et al., 20 Oct 2025).

7. Structural Features and Tightness of Weak Rate Bounds

The main determinants of the weak rate are:

  • Regularity of the coefficients, functional class of test observables, and stochastic integrability properties.
  • Geometric or analytic inequalities (Poincaré, weak-Poincaré, Log-Sobolev) controlling variance decay.
  • Tail properties of the invariant measure or potential.
  • The structure of the discretization scheme: explicit/implicit, projection methods, and adaptivity.

In most settings, weak convergence rates are strictly better than strong rates; for example, in many SPDEs, the weak rate is double the strong rate (Cai et al., 2019). However, in the absence of regularity or under roughness or degeneracy (e.g., in rough volatility models, degenerate diffusions, heavy-tailed targets), only sublinear or subgeometric (polynomial, stretched-exponential, logarithmic) weak rates are available.

Lower bound results demonstrate that, in many settings, the proven weak rates are sharp—reflecting both algorithmic and analytic limitations (e.g., Talay–Tubaro expansion for SDEs, information-theoretic lower bounds for optimization, tail-driven rates for MCMC) (Altmayer et al., 2016, Brown et al., 2024, Grimmer, 2021).

Table: Representative Weak Convergence Rate Bounds

Domain Setting / Model Weak Rate Bound Example Reference
Discretized SDE Log-Heston, ν>2\nu>2 O(Δ)O(\Delta) (Altmayer et al., 2016)
Rough Volatility Hurst HH O(Δ2H)O(\Delta^{2H}), O(ΔH+1/2)O(\Delta^{H+1/2}) (Bayer et al., 2022)
SPDE (Allen–Cahn) Additive noise, degree γ\gamma O(λNγ+τγ)O(\lambda_N^{-\gamma} + \tau^\gamma) (Cai et al., 2019)
Degenerate Diffusion Weak Poincaré, power tails O(exp(Ctω)), O((1+t)δ)O(\exp(-C t^\omega)),\ O((1+t)^{-\delta}) (Grothaus et al., 2017)
Inverse Problems Tikhonov reg., RKHS O(nmin(τ/d,1/2))O(n^{-\min(\tau/d,1/2)}) (Guastavino et al., 4 Dec 2025)
Adaptive MCMC Student-tνt_\nu target O(t(v+d2))O(t^{-(v+d-2)}) (Brown et al., 2024)
Probability Flow ODE Weak log-concave O(emT+h+ε)O(e^{-mT} + h + \varepsilon) (Kremling et al., 20 Oct 2025)
Multiscale SDE CLT McKean–Vlasov systems O(ϵ)O(\epsilon) for suitable test observables (Xiang et al., 20 May 2025)

These results indicate that under general circumstances, the precise weak convergence rate is dictated by the interplay between noise regularity or system degeneracy, the choice of test functions, and underlying analytic/geometric properties of the process or scheme.

References

Topic to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Weak Convergence Rate Bounds.