Continuous-Time Subgradient Trajectories
- Continuous-time subgradient trajectories are absolutely continuous curves that solve differential inclusions defined by the subdifferential of an objective function.
- They generalize classical gradient methods to nonsmooth and composite settings, ensuring convergence under conditions such as the KL property and Lyapunov descent.
- Their analysis connects discrete algorithms with continuous dynamics, offering insights into robust optimization applications including robust PCA, phase retrieval, and hierarchical minimization.
A continuous-time subgradient trajectory is an absolutely continuous curve in a finite- or infinite-dimensional space that solves a differential inclusion involving the (possibly nonsmooth, nonconvex) subdifferential of an objective function. These flows generalize classical gradient dynamics to nonsmooth or composite settings, form the theoretical backbone of modern first-order methods for robust and high-dimensional optimization, and have deep connections to geometric and variational analysis.
1. Definition and Mathematical Framework
Let be a locally Lipschitz function. The Clarke subdifferential at is defined as
where the Clarke directional derivative is
A (continuous-time) subgradient trajectory is any absolutely continuous curve that satisfies the differential inclusion
for almost every (Josz et al., 2023, Lewis et al., 2022, Cai et al., 15 Jan 2026).
In the convex or Hilbert space setting, with proper, closed, and convex, the inclusion generalizes to maximal monotone operator flows:
The subgradient flow provides a continuous counterpart to discrete-time subgradient methods and supplies the limiting behavior for first-order optimization algorithms under vanishing step sizes (Cai et al., 15 Jan 2026, Zhang et al., 2024).
2. Existence, Uniqueness, and Regularity
For proper, lower semicontinuous, bounded below, and primal-lower-nice functions, for every initial point , there exists a unique locally absolutely continuous curve satisfying the continuous-time subgradient inclusion, and the energy dissipation identity holds:
When is convex, the existence and uniqueness follow from classical monotone operator theory (Brézis, Rockafellar). For nonconvex, nonsmooth functions, primal-lower-nice or related regularity conditions are required to guarantee well-posed flows (Lewis et al., 2022).
3. Asymptotic Behavior, Stability, and Convergence
a. Lyapunov Descent and Critical Value Trapping
For functions locally Lipschitz, coercive, and tame (i.e., definable in an o-minimal structure), every subgradient trajectory is characterized by:
- nonincreasing and convergent to a critical value as ;
- Existence of a connected component of the critical set such that ;
- Once close to , remains in every neighborhood of after some finite time (Josz et al., 2023).
Key inequalities include:
and, by the Kurdyka–Łojasiewicz (KL) inequality,
in a neighborhood of each critical level .
Integration yields , enforcing trajectory convergence and vanishing velocity at infinity (Josz et al., 2023, Lewis et al., 2022, Cai et al., 15 Jan 2026).
b. KL Inequality and Identifiability
For proper, lower semicontinuous functions satisfying the KL property at a critical point, every subgradient trajectory converges:
After finite time, the trajectory is confined to an "identifiable" set, e.g., a manifold, and subsequently follows the projected smooth gradient dynamics on this set (Lewis et al., 2022).
c. Nonautonomous and Multiscale Generalizations
In the nonautonomous setting, the inclusion generalizes as
where evolves in time, possibly non-monotonically. Under Mosco-type or monotonicity assumptions on , trajectory convergence toward is established, where is the limit functional (Attouch et al., 2016).
Multiscale variants of the flow,
reflect hierarchical or viscosity-type minimization and exhibit weak (or in special settings, strong) convergence to solutions of a hierarchical minimization hierarchy (Attouch et al., 2016).
4. Pathological and Nonconvergent Dynamics
Lipschitz continuity of alone does not guarantee that continuous-time subgradient trajectories approach critical points. Explicit constructions exist where periodic bounded trajectories are confined away from all critical points:
- For well-designed in , it is possible to select a continuous subgradient vector field such that the associated ODE induces a bounded periodic orbit that remains disjoint from for all time (Daniilidis et al., 2019).
This demonstrates that neither convexity nor regularity can be dispensed with if one wishes to guarantee approachability of generalized stationary points. This pathology is present in both continuous and discrete time; in the discrete case, the corresponding subgradient iterates may also have non-diverging, non-converging bounded orbits under analogous constructions (Daniilidis et al., 2019).
5. Connection to Discrete Algorithms and Lyapunov Functions
Discrete subgradient algorithms often admit an interpretation as an inexact or time-discretized version of continuous subgradient flows. The key technical step in many analyses is to demonstrate that the discrete sequence remains close to some continuous-time trajectory, and to exploit the Lyapunov structure of the latter.
For instance, for locally Lipschitz, semialgebraic , the existence of uniform finite-length bounds for trajectories and their KL property enables the proof of convergence for discrete subgradient iterates with prescribed step-size regimes (e.g., ), provided all continuous subgradient trajectories of are bounded (Cai et al., 15 Jan 2026).
In decentralized or stochastic settings, one identifies a Lyapunov function (possibly including consensus and momentum components) that decreases along solutions of the limit dynamics. Under appropriate stability of the Lyapunov level set and boundedness assumptions, one concludes global convergence for all perturbed or consensus-coupled processes whose interpolations approximate the continuous-time behavior (Zhang et al., 2024).
6. Applications and Composite Extensions
Continuous-time subgradient dynamics underlie the analysis and design of algorithms for:
- Robust PCA, robust phase retrieval, robust matrix sensing, where the boundedness of subgradient flows is verified via analytic or geometric arguments depending on the subproblem structure (Cai et al., 15 Jan 2026).
- Proximal-gradient flows for objectives , analyzed via the ODE
which in the limit converges to classic subgradient flow. Exponential convergence is warranted under proximal-PL or KL-type conditions (Gokhale et al., 2024).
- Hierarchical minimization and selection principles in multiscale and nonautonomous systems (Attouch et al., 2016).
The continuous-time subgradient theory is thus fundamental to understanding the long-term and stability properties of first-order methods, both in convex and in broad classes of structured nonconvex optimization.
Key literature: (Josz et al., 2023, Daniilidis et al., 2019, Lewis et al., 2022, Cai et al., 15 Jan 2026, Zhang et al., 2024, Gokhale et al., 2024, Attouch et al., 2016).