Bregman Restricted Strong Convexity
- Bregman Restricted Strong Convexity is a framework that extends classical strong convexity by using Bregman divergence to measure curvature in sparse optimization settings.
- It ensures local convexity along sparse directions, which facilitates uniqueness of critical points and guarantees for ℓ₀-Bregman relaxation recovery.
- The approach adapts to non-quadratic losses, such as those in Poisson regression, by employing task-specific reference functions like smoothed Burg entropy.
Bregman Restricted Strong Convexity (B-RSC) is a generalization of the classical Restricted Strong Convexity (RSC) condition that plays a central role in modern nonconvex and sparse optimization, particularly in the analysis of -Bregman relaxations and variational methods involving non-Euclidean geometries. B-RSC quantifies the curvature of a function not in the standard metric but with respect to an adapted Bregman divergence, enabling effective analysis of problems where quadratic geometry is inadequate—such as generalized linear models or Kullback–Leibler data terms.
1. Formal Definition and Distinction from Classical RSC
Let , be proper, convex, and differentiable on . The Bregman divergence associated with is
and its symmetric version is
Definition (Bregman Restricted Strong Convexity, B-RSC): satisfies the Bregman Restricted Strong Convexity property with respect to over at sparsity level if there exists so that for all and all with , ,
If , , and B-RSC reduces to classical RSC:
In practice, B-RSC allows for non-quadratic, task-adapted geometries (e.g., via the Burg entropy for KL divergence) (Chirinos-Rodriguez et al., 15 Nov 2025).
2. Properties and Structural Consequences
Two key properties of B-RSC are monotonicity and restriction to sparse subspaces:
- Monotonicity: If satisfies B-RSC with constant , then for any , , B-RSC also holds with .
- Subspace Restriction: For any support with , define (where zero-pads). Then:
- is convex on ,
- (Chirinos-Rodriguez et al., 15 Nov 2025).
B-RSC thus ensures local strong convexity in -sparse directions relative to the geometry defined by . This generalizes the classical link between RSC and the lower Restricted Isometry Property (RIP) for quadratic data terms.
3. Motivation: Where Classical RSC Fails and B-RSC Succeeds
Classical RSC may be vacuous for non-quadratic losses. For instance, in Poisson regression with , , the global RSC constant since is not uniformly strongly convex as . In particular, fails RSC with .
B-RSC resolves this by matching the geometry: choosing (smoothed Burg entropy) renders the symmetric Bregman divergence nontrivial and positivity is restored under suitable lower-RIP conditions on , yielding . This allows effective landscape and recovery analysis even for Poisson/KL losses where classical RSC provides no information (Chirinos-Rodriguez et al., 15 Nov 2025).
4. Impact on Sparse Critical Points and Oracle Recovery
Within the -Bregman relaxation paradigm, B-RSC allows for rigorous statements regarding isolation and uniqueness of critical points.
- Isolation Theorem: Assume satisfies B-RSC with . If is a critical point of such that avoids specific bands tied to the subdifferential structure, for all , then any other critical point has . In particular, if , is the unique sparsest critical point.
- Uniqueness of Global Minimizer: Under the same B-RSC assumptions, if the sparsity penalty parameter exceeds , then is the unique global minimizer of (also of the original -penalized ), and all other critical points are at support distance greater than (Chirinos-Rodriguez et al., 15 Nov 2025).
B-RSC provides a mechanism that links a form of localized strong convexity to sparse recovery guarantees, encompassing both uniqueness and stability.
5. Oracle Solution Guarantees and “Safe Region” Design
With B-RSC (), where is the sparsity of the true underlying vector , one can guarantee that the oracle solution (the restricted minimizer over the true support) is the unique global minimizer if:
- There exists a “safe region” containing and the oracle minimizer, which avoids a “forbidden band” set in the coordinate directions;
- For all , certain subgradient bounds involving the measurement matrix columns and Bregman reference function hold;
- The parameter satisfies .
In practice, is formed so that all candidate minimizers stay strictly away from , with the size of determined by B-RSC: , where denotes the Bregman divergence along each active coordinate (Chirinos-Rodriguez et al., 15 Nov 2025).
6. Specialization: Quadratic and Poisson Data Terms
Least-Squares (LS) Setting
- Fidelity: , has Lipschitz gradient and is 1-strongly convex.
- Reference Function: , typically .
- B-RSC: For , B-RSC reduces to classical RSC, .
- Safe Region: .
- Oracle Recovery: One obtains explicit intervals for that often improve prior bounds (Chirinos-Rodriguez et al., 15 Nov 2025).
Poisson (KL) Data Setting
- Fidelity: , .
- Reference Function: , with tunable parameters.
- B-RSC: From Theorem 2.4 in (Chirinos-Rodriguez et al., 15 Nov 2025), a positive is achieved whenever satisfies a lower-RIP and is compact.
- Safe Region: .
- Oracle Recovery: Tight lower/upper bounds are derived, which are feasible in the high SNR regime, yielding nontrivial recovery guarantees unavailable in the RSC framework.
7. Relationship to Relative Bregman Strong Convexity and Algorithmic Implications
Related notions of strong convexity relative to Bregman divergences arise in optimization algorithm analysis. In the context of variance-reduced stochastic primal-dual splitting methods, strong convexity of the form
for a Legendre reference function (and similarly for ) ensures linear-primal dual gap contraction. The contraction rate depends explicitly on the Bregman relative strong convexity constant (Dung et al., 2021).
This formalism matches B-RSC in spirit when the restriction to sparse directions is taken, and when and are chosen to match the problem geometry. Notably, Bregman-based strong convexity is critical for linear convergence guarantees even in non-Euclidean settings, as shown by contraction constants and telescoping descent arguments in accelerated algorithms.
In summary, Bregman Restricted Strong Convexity extends strong convexity notions to non-Euclidean geometries and restricted (sparse) directions, enabling advanced landscape analysis and recovery theorems for nonconvex sparse optimization, particularly -Bregman relaxations under both quadratic and generalized linear models. It provides a unified framework for local convexity, isolation of minimizers, and explicit parameter regimes for exact oracle recovery beyond classical RSC applicability (Chirinos-Rodriguez et al., 15 Nov 2025, Dung et al., 2021).