Papers
Topics
Authors
Recent
Search
2000 character limit reached

Randomized Feasibility Algorithm with Polyak Steps

Updated 30 January 2026
  • The paper introduces a randomized feasibility algorithm that replaces full projection onto intersected constraints with tractable, sampled Polyak subgradient updates.
  • It employs adaptive, parameter-free step-size strategies to achieve linear convergence in strongly convex cases and optimal sublinear rates for general convex functions.
  • Empirical evaluations on QCQP and SVM tasks demonstrate that the method maintains computational efficiency and competitive performance without rigorous parameter tuning.

A randomized feasibility algorithm with Polyak steps is a class of iterative methods for constrained convex optimization where computationally tractable projections onto each individual constraint set are used instead of direct projection onto the intersection of all constraints. At each iteration, the algorithm randomly samples constraints and projects the current point towards feasibility using subgradient steps of Polyak type. Adaptive, problem-parameter-free step-size rules and sampled constraint selection enable linear or sublinear convergence rates according to the regularity of the objective function, while maintaining computational practicality when the full constraint projection is prohibitive (Chakraborty et al., 27 Jan 2026).

1. Problem Formulation and Notation

The central problem is of the form

minimizef(x)subject toxXY,\text{minimize} \quad f(x) \quad \text{subject to} \quad x \in X \cap Y,

where

  • f:RnRf: \mathbb{R}^n \to \mathbb{R} is convex (possibly strongly convex and/or smooth),
  • X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}, with each gig_i convex,
  • YRnY \subset \mathbb{R}^n is a simple closed convex set (such as a box or Euclidean ball).

Key notations:

  • \|\cdot\| is the Euclidean norm,
  • ΠY[z]\Pi_Y[z] denotes projection onto YY,
  • g+(x):=max{0,g(x)}g^+(x) := \max\{0, g(x)\},
  • dist(x,XY):=minyXYxy\text{dist}(x, X \cap Y) := \min_{y \in X \cap Y} \|x - y\|.

A global error bound assumption is used: there exists f:RnRf: \mathbb{R}^n \to \mathbb{R}0 and a sampling distribution f:RnRf: \mathbb{R}^n \to \mathbb{R}1 over f:RnRf: \mathbb{R}^n \to \mathbb{R}2 such that for all f:RnRf: \mathbb{R}^n \to \mathbb{R}3,

f:RnRf: \mathbb{R}^n \to \mathbb{R}4

2. Randomized Feasibility Algorithm with Polyak Steps

The algorithm performs a sequence of feasibility updates, each consisting of f:RnRf: \mathbb{R}^n \to \mathbb{R}5 substeps at iteration f:RnRf: \mathbb{R}^n \to \mathbb{R}6. Each feasibility substep involves:

  • Sampling a constraint index f:RnRf: \mathbb{R}^n \to \mathbb{R}7 uniformly,
  • Computing a subgradient f:RnRf: \mathbb{R}^n \to \mathbb{R}8,
  • Updating via the Polyak-type step: f:RnRf: \mathbb{R}^n \to \mathbb{R}9 where X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}0 is a parameter, and projection is onto X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}1.

After X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}2 such substeps, X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}3. This scheme avoids projection onto X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}4, replacing it with computationally tractable projections onto X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}5 and randomized selection of individual constraints.

Under the error-bound and bounded subgradient assumptions, the following hold:

  • Nonexpansiveness: For any feasible X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}6, X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}7.
  • Geometric decrease in infeasibility: X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}8 where X=i=1m{xRngi(x)0}X = \bigcap_{i=1}^m \{x \in \mathbb{R}^n \mid g_i(x) \le 0\}9.

3. Interleaved Objective Minimization and Feasibility Updates

The algorithm alternates or interleaves randomized feasibility updates with (sub)gradient steps for objective minimization. Two major cases are considered:

Strongly Convex, gig_i0-Smooth Objective

Assumptions:

  • gig_i1 has gig_i2-Lipschitz gradient,
  • gig_i3 is gig_i4-strongly convex.

Algorithm steps:

  1. Compute gig_i5,
  2. Update gig_i6 using the randomized feasibility algorithm with gig_i7.

Adaptive Polyak-type step size: gig_i8 where gig_i9 is a prescribed accuracy.

Weighted averaging is used: YRnY \subset \mathbb{R}^n0

Convex, Possibly Nonsmooth Objective: Distance-over-Weighted-Subgradients (DoWS)

Assumptions:

  • YRnY \subset \mathbb{R}^n1 is convex (possibly nondifferentiable),
  • YRnY \subset \mathbb{R}^n2 is convex and bounded with diameter YRnY \subset \mathbb{R}^n3.

For YRnY \subset \mathbb{R}^n4 iterations:

  • Maintain YRnY \subset \mathbb{R}^n5,
  • YRnY \subset \mathbb{R}^n6; YRnY \subset \mathbb{R}^n7,
  • YRnY \subset \mathbb{R}^n8,
  • Compute YRnY \subset \mathbb{R}^n9,
  • Randomized feasibility update as above.

A weighted average output \|\cdot\|0 minimizes \|\cdot\|1.

4. Convergence Guarantees and Theoretical Rates

Strongly Convex, Smooth Case

For adaptive stepsizes as above and exponential weighting,

\|\cdot\|2

after \|\cdot\|3 outer iterations, provided the mean reduction in infeasibility per iteration meets a prescribed threshold (Chakraborty et al., 27 Jan 2026).

Convex, Possibly Nonsmooth Case

After \|\cdot\|4 iterations using DoWS with feasibility, the output \|\cdot\|5 satisfies

\|\cdot\|6

with \begin{align*} A_1(T) &= \frac{2 D M_f}{\sqrt{T}}\left(\frac{D}{r}\right){\frac{2}{T}\ln(e D2/r2)},\ A_2(\tau) &= D M_f \max_{1\le k\le\tau} \mathbb{E}[ (1-q){N_k/2} ],\ A_3(T) &= \frac{D M_f}{T} \left(\frac{D}{r}\right){\frac{2}{T}\ln(e D2/r2)} \sum_{k=1}{\tau} \mathbb{E}[(1-q){N_k/2}], \end{align*} yielding the optimal \|\cdot\|7 rate as \|\cdot\|8 up to sampling-determined terms.

For unbounded \|\cdot\|9, a tamed (logarithmically adjusted) variant of the DoWS step-size ensures bounded iterates and the same ΠY[z]\Pi_Y[z]0 expected error rate up to constants that grow logarithmically in ΠY[z]\Pi_Y[z]1.

5. Sampling Distribution Regimes and Computational Properties

Performance and theoretical rates depend critically on the sampling distribution of the number of feasibility substeps ΠY[z]\Pi_Y[z]2 at each outer iteration. For common regimes:

  • Deterministic polynomial growth: ΠY[z]\Pi_Y[z]3 ensures that the sum ΠY[z]\Pi_Y[z]4 is uniformly bounded.
  • Poisson sampling: ΠY[z]\Pi_Y[z]5 with ΠY[z]\Pi_Y[z]6 yields ΠY[z]\Pi_Y[z]7, which decays polynomially in ΠY[z]\Pi_Y[z]8.
  • Binomial sampling: ΠY[z]\Pi_Y[z]9 with YY0 gives similar decay properties.

Sub-polynomial growth of YY1 suffices to make sampling-driven error negligible at polylogarithmic cost in total feasibility steps.

6. Empirical Evaluation: QCQP and SVM Applications

Simulations were conducted on two canonical classes of problems:

Quadratically Constrained Quadratic Programming (QCQP)

The problem: YY2 was tested in three regimes:

  • (a) Strongly convex YY3, known YY4,
  • (b) Strongly convex, unknown YY5,
  • (c) Convex YY6, unknown YY7.

Baselines included the Nedić et al subgradient-projection, Arrow–Hurwicz and Alt-GDA primal-dual schemes, ACVI (ADMM+log-barrier), and CVXPY interior-point.

Key observations:

  • Adaptive Polyak-step algorithm achieved linear convergence in (a), requiring no prior knowledge of strong convexity or smoothness parameters.
  • DoWS and T-DoWS performed competitively in (b), (c), attaining the expected YY8 rate slope.
  • ACVI provided the fastest infeasibility decay but required expensive tuning.

Support Vector Machine (SVM) Soft-Margin Classification

For the SVM problem

YY9

the UCI Banknote, Breast-Cancer, and MNIST 3-vs-5 datasets were used. Only DoWS/T-DoWS and primal-dual (Arrow–Hurwicz/Alt-GDA) baselines were compared due to convexity.

Results:

  • DoWS/T-DoWS schemes reduced objective and infeasibility rapidly;
  • Test-set misclassification rates were competitive with cross-validated primal-dual methods;
  • Methods required no parameter tuning.

7. Theoretical Significance and Practical Implications

Randomized feasibility algorithms with Polyak steps provide a rigorously justified, computation-efficient approach to large-scale constrained convex optimization where projection onto intersected constraints is intractable. Theoretical results guarantee:

  • Linear convergence to any prespecified tolerance for strongly convex, g+(x):=max{0,g(x)}g^+(x) := \max\{0, g(x)\}0-smooth g+(x):=max{0,g(x)}g^+(x) := \max\{0, g(x)\}1;
  • Optimal g+(x):=max{0,g(x)}g^+(x) := \max\{0, g(x)\}2 rates in the convex, potentially nonsmooth setting;
  • Bounded sampling-driven error without demanding hyperparameter tuning or explicit knowledge of problem parameters.

Empirical results indicate practical competitiveness against state-of-the-art first-order and primal-dual methods, particularly when problem structure or scale make conventional projection approaches prohibitively costly (Chakraborty et al., 27 Jan 2026).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Randomized Feasibility Algorithm with Polyak Steps.