Minimax-Optimal Halpern Scheme
- The scheme establishes fixed-point convergence through a classical Halpern anchoring method that attains an O(1/N^2) worst-case rate.
- It employs algebraic invariant theory (H-invariance) to characterize algorithms that achieve minimax optimality for nonexpansive and Lipschitz operators.
- Adaptive variants and splitting schemes extend its applicability to monotone inclusions, convex-concave minimax problems, and saddle-point optimization.
The minimax-optimal Halpern scheme refers to a class of fixed-point iterative algorithms for nonexpansive or Lipschitz operators that provably attain the optimal worst-case convergence rate for driving the fixed-point residual to zero. This paradigm generalizes Halpern’s classical anchoring approach, and exact characterizations exist both for nonexpansive maps in Hilbert/normed spaces and for operators relevant to monotone inclusions, variational inequalities, and convex-concave minimax optimization. The minimax-optimality is established via tight non-asymptotic bounds, algebraic invariant theory (H-invariance), and lower bounds showing that no first-order scheme can improve upon the obtained rates under black-box oracle models.
1. Halpern Iteration Fundamentals and Minimax Rate
Halpern’s iteration for finding with nonexpansive () in a normed space is anchored to an initial point and uses a sequence of weights : The optimal choice for minimax convergence in the nonexpansive () case is , yielding
and no deterministic first-order method can improve on the bound in this setting (Yoon et al., 18 Nov 2025).
In the general Lipschitz regime (0, 1), the minimax-optimal sequence 2 results from recursively minimizing a tight, quadratic upper bound on 3: 4 with the recursion for 5: 6 and residual (for 7): 8 (Bravo et al., 22 Jan 2026). This recursion is tight: for each 9, there exists a 0-Lipschitz 1 and initialization for which equality holds.
2. H-Invariance Theory and Complete Characterization
The exhaustive algebraic theory of minimax acceleration for nonexpansive fixed-point algorithms is provided via H-invariance (Yoon et al., 18 Nov 2025). Any algorithm admitting the lower-triangular "moment-mixing" form
2
is described by its H-matrix. The family of minimax-optimal algorithms is precisely those with H-invariants: 3 and nonnegative H-certificates 4 (unique solution to a linear-quadratic identity). Only methods with these invariants and certificates attain the guaranteed 5 rate. The classical Halpern method (OHM) and its H-dual are extremal cases.
3. Behavior under Contractive and Expansive Operators
For 6 (contractions), the minimax-optimal Halpern sequence transitions from a sublinear Halpern phase to geometric Banach–Picard iteration, as 7 reaches 8 and stays there (from the index 9 for which 0). This yields rapid geometric decay 1. As 2, the Halpern phase length diverges and recovers the 3 rate for nonexpansive maps.
For 4 (expansive), the sequence 5 remains strictly less than 6, and the residual converges to 7, matching the minimal displacement on bounded domains. The minimax scheme is purely Halpern throughout.
4. Potential-Based Analysis and Lower Bounds
The minimax-optimal rates are validated through tight potential-based analysis (Diakonikolas, 2020). For monotone inclusions 8 (with cocoercive 9), setting step-weight 0 and backtracking on the operator constant yields
1
The "potential" function is quadratic-minus-linear in the operator evaluations and telescopes over the run.
Lower bounds are established via reductions: for any first-order method, problems exist for which 2 is unimprovable (e.g., convex-concave saddle problems in 3 dimensions) (Diakonikolas, 2020). The Halpern scheme matches these bounds up to polylogarithmic factors and is parameter-free.
5. Splitting Schemes and Extensions
Halpern-type minimax algorithms have been extended to monotone splitting and saddle-point settings (Tran-Dinh et al., 2021). For 4 (maximally monotone 5, Lipschitz 6), Halpern–Popov variants give
7
where 8 is the forward–backward residual. Two splitting schemes are constructed:
- Extra-Anchored Gradient-Splitting (EAG-Split): two resolvent calls per iteration.
- Past-Extra-Anchored Gradient-Splitting (PEAG-Split): only one computation each of 9, 0, and 1.
Application to convex–concave minimax problems (2) leverages the Halpern scheme for the skew-gradient optimality operator, establishing 3 convergence of the gradient norm, matching known lower bounds.
6. Adaptive Variants and Practical Refinements
Adaptive Halpern schemes, inspired by minimax theoretical bounds, track empirical residuals 4 to compute step sizes 5 that can outperform the non-adaptive minimax bound in practical settings (Bravo et al., 22 Jan 2026). The update is
6
with 7. This process is universally better or equal compared to the minimax rates.
Extensions further cover unbounded domains (via displacement bounds based on fixed-point distance), affine maps (closed-form for optimal 8), and strongly monotone operators (with restarted Halpern strategies).
7. Comparison to Alternative Acceleration and Practical Implications
Nesterov’s acceleration for monotone inclusions yields 9 last-iterate rates unless extra structure (e.g., cocoercivity) is present. Anchored Extragradient schemes match the Halpern 0 rate but require twice as many operator calls. The minimax-optimal Halpern schemes uniquely combine single oracle usage, explicit algebraic invariant theory, and optimality under black-box oracle models and mere monotonicity/Lipschitz assumptions (Tran-Dinh et al., 2021, Yoon et al., 18 Nov 2025, Bravo et al., 22 Jan 2026).
Halpern-type minimax schemes unify approaches for splitting, inclusion, variational inequality, and convex–concave min–max, always matching fundamental lower bounds up to log-factors. The certificate-based H-invariance characterization catalogues every possible optimal algorithm in this class. These results constitute the canonical foundation for minimax-optimal first-order fixed-point algorithms.