Papers
Topics
Authors
Recent
Search
2000 character limit reached

Prox-Convex Functions Overview

Updated 18 January 2026
  • Prox-convex functions are a generalized form of convex functions that guarantee a unique, firmly nonexpansive proximity operator even in nonconvex settings.
  • They underpin various proximal algorithms, such as the proximal point method and splitting techniques, ensuring convergence and robust performance.
  • These functions bridge convex, weakly convex, and difference-of-convex paradigms and are vital in applications like image recovery and variational inequality analysis.

A prox-convex function is a generalized convexity notion applied to functions for which the proximity operator is single-valued, firmly nonexpansive, and admits suitable descent-type inequalities, often even in certain nonconvex settings. Prox-convexity subsumes convex functions and many weakly convex, quasiconvex, or difference-of-convex (DC) classes, but does not coincide with any of them. This notion is foundational to proximal point algorithms and their extensions to composite or nonconvex problems, enabling both theoretical guarantees and practical algorithms that go beyond the classical convex regime. Recent research elaborates diverse definitions, properties, structural results, algorithmic templates, and practical scenarios in which prox-convexity is essential.

1. Formal Definitions and Basic Properties

Prox-convexity is defined via the behavior of the proximity operator. For a proper, lower semicontinuous function f:RnR{+}f: \mathbb{R}^n \to \mathbb{R} \cup \{+\infty\} and a closed set KK, ff is prox-convex on KK with constant α>0\alpha > 0 if for every zKz \in K, the subproblem

prox ⁣f,K(z):=argminxK{f(x)+12αxz2}\operatorname{prox}_{\!f,K}(z) := \arg\min_{x \in K} \{ f(x) + \tfrac{1}{2\alpha} \|x - z\|^2 \}

has a unique solution TT, and for all xKx \in K, the inequality

f(x)f(T)α(zT)(xT)f(x) - f(T) \ge \alpha \, (z-T)^\top (x-T)

holds (Grad et al., 2021). This captures both existence and a firm nonexpansivity property for the prox mapping. Convex functions furnish prox-convexity automatically with α=1\alpha=1; prox-convexity extends to weakly convex (cc-weakly convex) functions with α<1/c\alpha < 1/c (Davis et al., 2019).

Prox-convex functions strictly include some nonconvex cases, e.g., f(x)=x2xf(x)=-x^2-x on [0,1][0,1] (Grad et al., 2021). The proximity operator for a prox-convex function is always single-valued, and the associated map is firmly nonexpansive:

T(z)T(w)2+(zT(z))(wT(w))2zw2\|T(z)-T(w)\|^2 + \|(z-T(z)) - (w-T(w))\|^2 \le \|z-w\|^2

for all z,wKz, w \in K.

2. Connections to Generalized Convexity Paradigms

Prox-convexity relates to—but does not coincide with—several important generalized convexity classes:

  • Convex functions: Every convex function is prox-convex with modulus 1.
  • Weakly convex functions: If f(x)+c2x2f(x) + \frac{c}{2}\|x\|^2 is convex (cc-weakly convex), then ff is prox-convex for any α<1/c\alpha < 1/c (Davis et al., 2019, Grad et al., 2021).
  • Quasiconvex and DC functions: Many (strongly) quasiconvex and difference-of-convex functions are prox-convex, but the inclusion is not bidirectional (Grad et al., 2021), as certain counterexamples demonstrate.
  • Semi-algebraic and o-minimal losses: Prox-convex models encompass semi-algebraic structures, e.g., 1\ell_1 penalties, ReLU activations, nuclear norm surrogates, and truncated quadratic clustering losses (Davis et al., 2019).

A summary of containment relationships:

Class Contains prox-convex Subset of prox-convex Intersection nontrivial
Convex Yes Yes Yes
Weakly convex Yes Yes Yes
DC functions No No Yes
Strongly quasiconvex No No Yes

3. Proximal Operators and Decomposition

For f,gΓ0(H)f,g \in \Gamma_0(H) on a Hilbert space, the proximal operator of their sum satisfies a decomposition:

proxf+g=proxfproxgf,\operatorname{prox}_{f+g} = \operatorname{prox}_f \circ \operatorname{prox}_g^f,

where proxgf(x)\operatorname{prox}_g^f(x) is defined as the unique yy such that xyg(proxf(y))x-y \in \partial g(\operatorname{prox}_f(y)) (Adly et al., 2017). This naturally generalizes the classical Douglas-Rachford splitting, aids variational sensitivity analysis, and gives tractable fixed-point algorithms for sums of prox-convex functions.

For composite functions F(x)=g(x)+h(C(x))+s(R(x))F(x) = g(x) + h(C(x)) + s(R(x)) (with g,hg,h convex, C,sC,s smooth, components rir_i convex), the prox-convex approach forms a convex subproblem at each iteration by linearizing only the smooth maps and keeping convex terms exact, with strong convexification via a metric Qk=μkI+Hk+Q_k = \mu_k I + H_k^+ (Uzun et al., 22 Dec 2025). This structure enables robust global convergence and local Q-linear contraction.

4. Algorithmic Frameworks for Prox-Convex Minimization

Classical and modern proximal algorithms leverage prox-convexity for both convex and certain nonconvex objectives. Key schemes include:

  • Proximal Point Algorithm (PPA): Iteratively solves xk+1=proxf(xk)x_{k+1} = \operatorname{prox}_f(x_k). For prox-convex ff, PPA yields monotonic decrease, bounded iterates in sublevel sets, and O(1/k)O(1/k) Moreau gap rate (Grad et al., 2021). Convergence to stationary points is guaranteed on closed convex sets.
  • Splitting PPA: For f=ifif = \sum_i f_i (each prox-convex, Lipschitz), splitting PPA evaluates individual proxβkfi\operatorname{prox}_{\beta_k f_i} steps in cyclic or random order (Brito et al., 11 Jan 2026). Both deterministic and stochastic variants admit global convergence; the stochastic variant exploits supermartingale arguments for almost sure convergence.
  • Composite/Relaxed Proximal Algorithms: For F=g+hC+sRF=g+h\circ C+s\circ R, the prox-convex step (typically,

xk+1=argminxFQk(x;xk)x_{k+1} = \arg\min_x F_{Q_k}(x;x_k)

with FQk(x;xk)=F(x;xk)+12xxkQk2F_{Q_k}(x;x_k) = F(x;x_k) + \frac{1}{2} \|x-x_k\|_{Q_k}^2) preserves all convex structure and achieves O(ε2)O(\varepsilon^{-2}) complexity for prox-gradient residuals, with Q-linear rates under local error bounds (Uzun et al., 22 Dec 2025).

  • Douglas-Rachford and other fully proximal splitting: These methods employ only proximal activations, which are now feasible with closed-form prox operators established for many smooth convex penalties (Combettes et al., 2018).

5. Interpretation of Prox-Convexity in Applications and Composite Optimization

Prox-convexity is instrumental in numerous domains:

  • Weakly Convex/Nonsmooth Optimization: Proximal algorithms can escape strict saddles and converge to local minima for weakly convex functions satisfying the strict-saddle property, particularly in problems with semi-algebraic loss landscapes (Davis et al., 2019). This underpins robust minimization in modern machine learning and signal processing.
  • Image Recovery and Structured Inverse Problems: Fully proximal activation, enabled by closed-form prox operators for smooth terms, substantially accelerates convergence in image deconvolution, reconstruction, interpolation, and inconsistent feasibility relaxation. Practical comparisons consistently show that the fully proximal strategy outperforms mixed gradient/proximal approaches (Combettes et al., 2018).
  • Variational Inequality Sensitivity: The decomposition formula for the sum of two convex (or prox-convex) functions computes directional derivatives of solution maps for linear variational inequalities, integrating seamlessly with established sensitivity analysis frameworks (Adly et al., 2017).

6. Comparison Principles, Determination, and Lipschitz Characterization

Proximal mapping properties encode substantial information about the underlying function:

  • Comparison principles: If proxf(x)x0proxg(x)x0\|\operatorname{prox}_f(x) - x_0\| \le \|\operatorname{prox}_g(x) - x_0\| for all xx, then g(x)g(x0)f(x)f(x0)g(x) - g(x_0) \le f(x) - f(x_0) pointwise (Vilches, 2020).
  • Determination by Proximal Norm: The pointwise norm proxf(x)\|\operatorname{prox}_f(x)\| uniquely determines a convex ff: if two convex functions have matching norms for all xx, they differ only by an additive constant. This equivalence extends through minimal-norm subgradients and Moreau envelopes (Vilches, 2020).
  • Lipschitz Characterization: Convex ff is \ell-Lipschitz iff xproxf(x+y)y\|x\|-\ell \le \|\operatorname{prox}_f(x+y)-y\| for all x,yx,y (Vilches, 2020).

These results facilitate function identification and classification purely through observed proximal dynamics.

7. Extensions in Dual Averaging and Prox-Functions

Dual averaging schemes commonly require a prox-function hh that is strongly convex; recent work relaxes this, establishing O(1/k)O(1/k) rates under prox-convex-like assumptions:

  • Prox-Compact + Domain Inclusion: Strong convexity on the compact region containing iterates suffices, provided certain primal/dual domain inclusion conditions hold (Zhao, 4 Apr 2025).
  • Dual-Monotonicity and Open Domain: When the Fenchel dual domain of hh^* is open, dual monotonicity can be enforced for convergence (Zhao, 4 Apr 2025).
  • Function Classes: Many entropic, barrier, and "indicator + barrier" combinations meet these requirements, broadening the set of viable regularizers in machine learning and structured optimization.

8. Discrete-Choice Prox-Functions and Specialized Models

Discrete-choice geometry induces a specialized class of prox-functions on the simplex, derived as convex conjugates of surplus (log-sum) functions arising in random utility models (Müller et al., 2019). These prox-functions are strongly convex with respect to 1\ell^1 norm (by Legendre duality) and admit closed-form mirror descent updates with probabilistic interpretation in terms of choice probabilities. Strong convexity parameters are computed explicitly for generalized extreme value and nested logit models. Such constructions enable dual-averaging schemes with O(1/k)O(1/\sqrt{k}) convergence in the duality gap, naturally fitting economic application contexts.

References

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Prox-Convex Functions.