Factorized Robust Signal Recovery
- Factorized Robust Signal Recovery is an approach that decomposes signals into factors to robustly recover underlying structures even in nonconvex, nonsmooth settings.
- The framework employs continuous-time subgradient trajectories and differential inclusions to establish convergence and identify active manifolds in complex optimization landscapes.
- It underpins robust algorithms used in applications such as robust PCA, phase retrieval, and neural network training, enabling effective recovery of corrupted signals.
A continuous-time subgradient trajectory is an absolutely continuous curve that evolves according to a differential inclusion driven by the (typically nonconvex, possibly nonsmooth) subdifferential of an objective function. Such trajectories serve as a mathematical bridge between discrete optimization algorithms, such as the subgradient method, and the dynamical systems perspective, offering tools to analyze asymptotic convergence, manifold identification, and pathological behaviors in nonsmooth, nonconvex, or time-dependent settings. The formalism accommodates both static and nonautonomous (time-varying) optimization, tracks critical-point convergence via Lyapunov and Kurdyka–Łojasiewicz (KL) structures, and underpins the performance of first-order methods in largescale signal recovery, machine learning, and variational analysis.
1. Differential Inclusion Formalism and Subdifferentials
Let be locally Lipschitz. The Clarke subdifferential at is defined as
where denotes the Clarke directional derivative. A continuous-time subgradient trajectory is an absolutely continuous curve satisfying
for almost every . This is a differential inclusion generalizing the notion of gradient flow to nonsmooth, possibly nonconvex settings (Josz et al., 2023, Lewis et al., 2022, Cai et al., 15 Jan 2026).
For general maximal monotone operators in Hilbert spaces, the dynamical system
arises, with particular interest in the case where for a proper, closed, convex function (Attouch et al., 2016).
2. Existence, Uniqueness, and Regularity
If is proper, lower semicontinuous, bounded below, and primal-lower-nice (a generalization covering convex, weakly convex, and strongly amenable composite functions), any initial point yields a unique, locally absolutely continuous trajectory satisfying the subgradient differential inclusion (Lewis et al., 2022). These trajectories have finite energy, quantified by
In the convex case, classical results by Brézis and Marcellin–Thibault establish this for general maximal monotone evolutions (Lewis et al., 2022, Attouch et al., 2016).
In time-varying (nonautonomous) settings, for a family of proper, closed, convex functions , global solutions exist for the nonautonomous inclusion (NAGI)
with convergence properties determined by Mosco-type assumptions and energetical inequalities (Attouch et al., 2016).
3. Global Convergence and Lyapunov Structure
For a broad class of objectives—locally Lipschitz, coercive, and definable (tame) in an o-minimal structure (e.g., semialgebraic)—every continuous-time subgradient trajectory converges to a neighborhood of a single connected component of the set of critical points . Specifically:
- The objective is nonincreasing and absorbs to a critical value as .
- The distance to a connected component of critical points vanishes: .
- After some finite time, the trajectory remains in every pre-specified neighborhood of .
These properties exploit the Kurdyka–Łojasiewicz (KL) inequality, with desingularizing functions encoding how the subdifferential norm lower-bounds the function-value gap (Josz et al., 2023, Zhang et al., 2024). Along any such trajectory,
leading to function values converging monotonically, and, under KL-type conditions, the entire trajectory has finite length: Discrete-time subgradient methods can be shown to track these flows over finite intervals and inherit their global convergence guarantees under suitable step-size regimes (Josz et al., 2023, Cai et al., 15 Jan 2026, Zhang et al., 2024).
4. Pathological and Exceptional Dynamics
Despite the general convergence guarantees under coercivity, definability, or suitable KL properties, continuous-time subgradient flows may admit pathological behaviors for merely Lipschitz functions (Daniilidis et al., 2019). Explicit examples with Lipschitz in exist where:
- Trajectories are periodic and remain bounded,
- The critical set (where ) is never visited by the trajectory,
- The objective is not necessarily monotone along the flow.
The mechanism exploits the freedom within the set-valued Clarke subdifferential to engineer ODE selections whose solutions are classical rotations, avoiding all critical points. These constructions demonstrate that, absent regularity (tameness, convexity, KL conditions), Lipschitz continuity of is insufficient for subgradient convergence or Lyapunov decrease (Daniilidis et al., 2019). Analogous phenomena can occur in discrete-time subgradient methods with nonsummable step-sizes.
5. Extensions: Nonautonomous, Multiscale, and Proximal-Gradient Flows
Nonautonomous Gradient Systems
For time-dependent convex objectives , convergence of solutions to the minimizer set of the limiting function requires Mosco-type conditions and control of the Brézis–Haraux excess function
Appropriate summability conditions yield weak convergence to the minimizer set, while strong minima enable pointwise convergence (Attouch et al., 2016).
Multiscale and Hierarchical Minimization
Multiscale systems of the form
encode hierarchical minimization, selecting minimizers of over the set . Viscosity-type inclusions select regularized (e.g., minimal-norm) solutions (Attouch et al., 2016).
Proximal-Gradient Dynamics
For composite objectives , where is differentiable convex and is convex (possibly nonsmooth), proximal-gradient flows are captured by
In the limit , this recovers the subgradient flow . For satisfying a proximal-PL or KL-type inequality,
one obtains exponential convergence of to the minimal value (Gokhale et al., 2024).
6. Identifiability, Partial Smoothness, and Manifold Identification
Identifiability and partial smoothness predict that subgradient trajectories may, after finite time, be confined to a lower-dimensional manifold (active set) on which the objective is smooth. Under metric identifiability (no vanishing subgradients when approaching from outside the manifold), and regularity hypotheses (primal-lower-niceness, local smoothness), subgradient flows are shown to enter the manifold in finite time and subsequently evolve according to projected smooth dynamics (Lewis et al., 2022). For example, for
all subgradient curves converge to , but after some finite time, trajectories enter the parabola and continue as smooth flows.
7. Applications and Impact on Algorithmic Analysis
Continuous-time subgradient trajectories underpin the analysis of convergence and avoidance of spurious critical points for discrete-time first-order methods applied to robust PCA, phase retrieval, decentralized stochastic optimization, and neural network training (Zhang et al., 2024, Cai et al., 15 Jan 2026, Gokhale et al., 2024). For semialgebraic and tame objectives, the boundedness and finite-length properties of subgradient flows ensure that for sufficiently small step-sizes, discrete iterates are trapped in neighborhoods of critical sets and converge globally, even in nonconvex, nonsmooth scenarios (Josz et al., 2023, Cai et al., 15 Jan 2026). Exceptions and cautionary examples limit this paradigm to settings with sufficient regularity, as evidenced by pathological constructions (Daniilidis et al., 2019).
The continuous-time perspective is thus a unifying analytic tool throughout modern nonsmooth optimization, geometrically informed analysis, and decentralized consensus, enabling precise statements about global dynamics, invariant sets, convergence rates, and the structure of limiting behaviors across a wide range of mathematical and algorithmic contexts.