Papers
Topics
Authors
Recent
Search
2000 character limit reached

Proximal Gradient Descent: Theory & Applications

Updated 28 January 2026
  • Proximal Gradient Descent is a method for composite optimization, combining a gradient step on smooth functions with a proximal step for nonsmooth regularizers.
  • It enables practical solutions for high-dimensional problems in sparse regression, matrix factorization, and signal recovery through adaptive and accelerated variants.
  • Recent advances demonstrate rigorous convergence guarantees, distributed implementations, and integration with learned regularizers for enhanced performance.

Proximal gradient descent is a foundational algorithmic framework in convex and nonconvex optimization for minimizing composite objective functions of the form

minxRnF(x):=f(x)+g(x),\min_{x \in \mathbb{R}^n} F(x) := f(x) + g(x),

where ff is typically smooth (i.e., differentiable with Lipschitz-continuous gradient) and gg is convex but potentially nonsmooth. The method iteratively combines explicit gradient descent on ff with implicit regularization via the proximal operator of gg, which is essential in high-dimensional settings with structured priors such as sparsity or low-rank. This approach undergirds algorithms for a wide array of regularized regression, signal recovery, matrix factorization, and machine learning applications, and recent work has established rigorous convergence rates, acceleration phenomena, distributed variants, and integration with learned regularizers.

1. Mathematical Formulation and Core Algorithms

Let f:RnRf: \mathbb{R}^n \to \mathbb{R} be convex and LL-smooth (f(x)f(y)Lxy\|\nabla f(x) - \nabla f(y)\| \leq L\|x - y\|), and g:Rn(,+]g: \mathbb{R}^n \to (-\infty, +\infty] be proper, closed, and convex (possibly nonsmooth). The proximal operator for gg at vv with parameter α>0\alpha>0 is defined as

proxαg(v)=argminx{g(x)+12αxv2}.\operatorname{prox}_{\alpha g}(v) = \arg\min_x \left\{ g(x) + \frac{1}{2\alpha} \|x - v\|^2 \right\}.

The proximal gradient iteration is

xk+1=proxαg(xkαf(xk)).x^{k+1} = \operatorname{prox}_{\alpha g}\left(x^k - \alpha \nabla f(x^k)\right).

For regularizers like the 1\ell^1 norm, this reduces to soft-thresholding; for nuclear norm penalties, the proximal step is singular value thresholding (Nikolovski et al., 2024, Zhang et al., 2015).

Accelerated schemes (e.g., FISTA) and inertial variants (proximal heavy-ball, β\beta-momentum) are formulated by supplementing the update with momentum terms or potential-function-based modifications, providing faster convergence in many cases (Sun et al., 2018, Chen et al., 2022, Bok et al., 2024).

2. Theoretical Guarantees and Rate Results

Proximal gradient descent enjoys the following canonical convergence properties:

Gα(x)=1α[xproxαg(xαf(x))],G_\alpha(x) = \frac{1}{\alpha}\left[x - \operatorname{prox}_{\alpha g}(x - \alpha \nabla f(x))\right],

contracts exactly by a factor p(α)=max{1Lα,1μα}p(\alpha) = \max\{|1 - L\alpha|, |1 - \mu\alpha|\} (Chen et al., 2022, Zhang et al., 2019). With optimal step size α=2/(L+μ)\alpha = 2/(L+\mu), this yields the tightest linear constant.

  • Under the Polyak–Łojasiewicz (PL) inequality, an improved objective convergence rate is attainable—specifically, (1ηα)/(1+ηα)(1-\eta\alpha)/(1+\eta\alpha) with appropriate η\eta (Zhang et al., 2019).

Recent work constructs potential-function frameworks leveraging norm-monotonicity of Gα(x)G_\alpha(x) and refined descent lemmas, yielding tight O(1/k)O(1/k) and accelerated O(1/k2)O(1/k^2) (function value), O(1/k3)O(1/k^3) (mapping norm) rates for composite problems (Chen et al., 2022).

Adaptive step-size rules in ProxGD, driven by observed local gradient differences, permit larger per-iteration steps and require only local Lipschitzness, not a global LL (Malitsky et al., 2023, Nikolovski et al., 2024).

3. Variants: Acceleration, Momentum, and Adaptive Step Sizes

Several lines of research extend proximal gradient descent in the following directions:

  • Momentum and inertia: The proximal inertial gradient descent (PIGD) update,

xk+1=proxαkg(xkαkf(xk)+βk(xkxk1)),x^{k+1} = \operatorname{prox}_{\alpha_k g}(x^k - \alpha_k \nabla f(x^k) + \beta_k(x^k - x^{k-1})),

achieves non-ergodic O(1/k)O(1/k) rates for convex problems with constant momentum (under coercivity), and linear rates under error-bound conditions (Sun et al., 2018).

  • Accelerated step schedules: The "silver" stepsize schedule, based on a quasifractal pattern related to the silver ratio, accelerates the rate of vanilla PGD for smooth convex ff from O(ε1)O(\varepsilon^{-1}) iterations to O(ε0.7864)O(\varepsilon^{-0.7864}) iterations without use of momentum or extrapolation. Under strong convexity, this generalizes to O(κ0.7864log(1/ε))O(\kappa^{0.7864} \log(1/\varepsilon)) (Bok et al., 2024).
  • Adaptive step size: Step sizes estimated from local curvature, αk=1/Lk\alpha_k = 1/L_k with LkL_k computed from gradient differences, allow for more aggressive updates and maintain theoretical O(1/k)O(1/k) rates without global Lipschitz constants (Malitsky et al., 2023, Nikolovski et al., 2024).

4. Distributed, Stochastic, and Variance-Reduced Proximal Schemes

For large-scale and distributed optimization, proximal schemes have been extended by:

  • Stochastic PGD: Replacing full gradients with stochastic (possibly minibatch) estimators enables scalable optimization for large data. Standard Prox-SGD achieves O(1/T)O(1/\sqrt{T}) (convex) or O(1/T)O(1/T) (strongly convex) convergence (Zhang et al., 2015).
  • Variance reduction: Epoch-based schemes (SVRG, SAGA) incorporated into proximal stochastic updates yield linear convergence guarantees for empirical risk minimization (Huo et al., 2016). The DAP-SVRG algorithm achieves linear convergence for strongly convex problems by combining asynchronous variance-reduction, worker-side proximal steps, and elementwise server updates.
  • Decoupled asynchronous variants: To minimize bottlenecks, DAP-SGD and DAP-SVRG offload proximal operator computations to worker nodes, permitting highly parallel, lock-free updates. The master only aggregates elementwise corrections, allowing near-linear speedup with the number of workers and supporting composite regularizers such as group 2\ell_2, nuclear norm, and fused lasso (Li et al., 2016, Huo et al., 2016).
  • Low-rank and structure-exploiting approaches: For high-dimensional matrix problems, stochastic proximal updates leverage low-rank sketches for the gradient, yielding space complexity O(m+n)O(m+n) rather than O(mn)O(mn) and convergence rates O((logT)/T)O((\log T)/\sqrt{T}) (convex) and O((logT)/T)O((\log T)/T) (strongly convex) (Zhang et al., 2015).

5. Extensions: Bregman Geometry, Online Optimization, and Plug-and-Play

  • Bregman Proximal Gradient Descent (BPGD): PGD can be generalized to problems endowed with a geometry induced by a strictly convex function hh via Bregman divergence Dh(x,y)=h(x)h(y)h(y),xyD_h(x, y) = h(x) - h(y) - \langle \nabla h(y), x - y \rangle. When ff is relatively smooth w.r.t hh, BPGD maintains descent with O(1/k)O(1/k) rates, and, under relative strong convexity or Bregman-PL, achieves global linear convergence. Multilevel Bregman schemes (ML-BPGD) exploit hierarchical discretizations and accelerate convergence for imaging and inverse problems (Elshiaty et al., 4 Jun 2025).
  • Online and inexact proximal updates: In streaming or adversarial dynamic settings, proximal OGD tracks a shifting composite optimum and admits dynamic regret bounds scaling with the path length of the optimum and cumulative gradient error (Dixit et al., 2018).
  • Plug-and-Play and learned regularization: Proximal-gradient algorithms lend themselves to integration with learned proximal maps (e.g., deep denoisers). Plug-and-play (PnP) methods replace the analytic proximal operator with a pretrained denoiser, requiring the denoiser to be proximal of some (possibly nonconvex, weakly convex) functional. Convergence can be recovered by introducing relaxation parameters in the update and controlled step size, with sufficient conditions on the regularization and data-fidelity weights (Hurault et al., 2023, Luo et al., 2022, Mardani et al., 2018).

6. Practical Implementation and Applications

  • Implementation: Each iteration requires one gradient evaluation and one proximal step, both of which can be computed efficiently for many standard regularizers (soft-thresholding for 1\ell^1, groupwise shrinkage, or singular value thresholding) (Nikolovski et al., 2024). Adaptive and acceleration schemes require minor bookkeeping or curvature estimation. Distributed and asynchronous execution (DAP-type algorithms) permit scalable training over large data and models with minimal master bottleneck.
  • Applications: Proximal gradient descent and its variants are core solvers for compressed sensing, sparse regression, low-rank matrix recovery, robust PCA, penalized multi-block CCA, and large-scale empirical risk minimization (Guan, 2022, Zhang et al., 2015, Nikolovski et al., 2024, Mardani et al., 2018, Luo et al., 2022).
  • Empirical findings: Variable step-size and adaptive PGD frequently reduce wall time and iteration counts by 40–50% relative to conservative fixed-step methods (Malitsky et al., 2023, Nikolovski et al., 2024). Nonlinear learned proximals in imaging and inverse problems yield substantial empirical gains, e.g., $3$ dB PSNR improvement in MRI reconstruction, over traditional 1\ell^1 regularization (Luo et al., 2022).

Proximal gradient descent interpolates between pure first-order methods for smooth losses and non-smooth minimization by subgradients or projections. Its key distinctions include:

Proximal gradient descent is thus the central algorithmic primitive for composite regularized optimization, underpinning scalable, robust, and adaptive solutions across theory and modern applications. The extensive theoretical and algorithmic toolkit guarantees performance across diverse regimes—accelerated, distributed, stochastic, adaptive—supported by rigorous and illustrative work in the literature (Nikolovski et al., 2024, Bok et al., 2024, Zhang et al., 2015, Nikolovski et al., 2024, Malitsky et al., 2023, Chen et al., 2022, Sun et al., 2018, Li et al., 2016, Huo et al., 2016, Elshiaty et al., 4 Jun 2025, Zhang et al., 2019, Guan, 2022, Hurault et al., 2023, Luo et al., 2022, Dixit et al., 2018, Mardani et al., 2018).

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Proximal Gradient Descent.