Papers
Topics
Authors
Recent
Search
2000 character limit reached

Subgame Perfect Proximal Point Algorithm

Updated 18 November 2025
  • SPPPA is a first-order method for nonsmooth convex optimization that uses a prox-oracle and history-dependent SOCP planning to achieve subgame perfection.
  • The algorithm dynamically re-optimizes its continuation strategy at each iteration, ensuring improved worst-case performance beyond classical minimax guarantees.
  • A limited-memory variant reduces computational cost by focusing on recent iterates, enabling efficient updates via Gram matrices and interior-point methods.

The Subgame Perfect Proximal Point Algorithm (SPPPA) is a first-order method designed for nonsmooth convex optimization utilizing a proximal operator oracle. SPPPA achieves the game-theoretic notion of subgame perfection, a property that requires not only minimax-optimality over the original problem class but also optimality on every subclass revealed through the observed information at each step. At every iteration, SPPPA dynamically re-optimizes its continuation strategy by solving a small, history-dependent second-order cone program (SOCP), which yields performance bounds that are guaranteed never to be worse and are typically strictly better than minimax-optimal guarantees for the original class (Grimmer et al., 17 Nov 2025).

1. Problem Setting and Oracle Model

The target is unconstrained convex minimization: minxRdf(x)\min_{x\in\mathbb{R}^d} f(x) where f:Rd(,+]f:\mathbb{R}^d\to(-\infty,+\infty] is closed, proper, convex, and attains its minimum at some yy_\star. The algorithm is initialized at x0x_0, and progress is evaluated by normalized suboptimality: f(yN)f12x0y2\frac{f(y_N)-f_\star}{\tfrac12\|x_0-y_\star\|^2} The model assumes access to a “prox-oracle of accuracy one”: for any query xx, the oracle returns

y=proxL,f(x)=argminu{f(u)+L2ux2},g=L(xy)f(y),f(y)y = \mathrm{prox}_{L,f}(x) = \arg\min_u \bigl\{f(u) + \tfrac{L}{2}\|u-x\|^2\bigr\},\quad g = L(x-y) \in \partial f(y),\quad f(y)

where {L0,L1,...,LN1}\{L_0, L_1, ..., L_{N-1}\} is a predefined sequence of positive proximal parameters.

Algorithms in this class, denoted A\mathcal{A}, are deterministic and at each step query the prox-oracle, then select their next iterate within the affine span generated by past subgradients: xnx0+span{g0,,gn1}x_n \in x_0 + \mathrm{span}\,\{g_0, \ldots, g_{n-1}\} with this selection depending solely on previously observed history {(yi,fi,gi,xi)}i=0n1\{(y_i, f_i, g_i, x_i)\}_{i=0}^{n-1}.

2. The Subgame Perfect Proximal Point Algorithm: Structure and Update

SPPPA generalizes the Optimized Proximal Point Algorithm (OPPA) by dynamically re-optimizing its proof of progress at each iteration. While OPPA proceeds by a closed-form recurrence, SPPPA, at each iteration nn, observes the current history

Hn1={(yi,fi,gi,xi,zi+1,τi)}i=0n1\mathcal{H}_{n-1} = \{(y_i, f_i, g_i, x_i, z_{i+1}, \tau_i)\}_{i=0}^{n-1}

and constructs matrices and vectors encoding this history:

  • Select the index mm with minimal observed value fm=min{f0,...,fn1}f_m = \min\{f_0,...,f_{n-1}\}
  • Form Z=[z1x0  znx0]Z = [z_1-x_0\ \cdots\ z_n-x_0], G=[g0  gn1]G = [g_0\ \cdots\ g_{n-1}] in Rd×n\mathbb{R}^{d \times n}
  • Vectors: τ=(τ0,,τn1)\tau = (\tau_0,\dots,\tau_{n-1})^\top, f=(f0,,fn1)f = (f_0,\dots,f_{n-1})^\top
  • Auxiliary vectors: ai=12zi+1x02+τi(fifm),bi=(figi,yix0)fma_i = \tfrac12\|z_{i+1}-x_0\|^2 + \tau_i(f_i - f_m),\quad b_i = (f_i - \langle g_i, y_i - x_0 \rangle) - f_m

The next step is determined by solving the following SOCP in non-negative variables μ,λRn\mu, \lambda_\star \in \mathbb{R}^n: maximize τ,μ+1λ\text{maximize }\, \langle\tau, \mu\rangle + \mathbf{1}^\top \lambda_\star

subject to 12ZμGλ2μ,a+λ,b\text{subject to }\, \tfrac12\|Z\mu - G\lambda_\star\|^2 \le \langle\mu,a \rangle + \langle \lambda_\star, b \rangle

Denote optima by (μ,λ)(\mu^\star, \lambda_\star^\star), then

z=x0+ZμGλ,τ=τ,μ+1λz' = x_0 + Z \mu^\star - G \lambda_\star^\star,\quad \tau' = \langle \tau, \mu^\star \rangle + \mathbf{1}^\top \lambda_\star^\star

With (z,τ)(z',\tau'), update as

τn=τ+1Ln(1+1+2Lnτ)\tau_n = \tau' + \frac{1}{L_n}\bigl( 1 + \sqrt{1 + 2L_n \tau'} \bigr)

xn=ττnym+τnττnzx_n = \frac{\tau'}{\tau_n} y_m + \frac{\tau_n-\tau'}{\tau_n} z'

yn=proxLn,f(xn),gn=Ln(xnyn),zn+1=z(τnτ)gny_n = \mathrm{prox}_{L_n, f}(x_n),\quad g_n = L_n(x_n - y_n),\quad z_{n+1} = z' - (\tau_n - \tau') g_n

3. Subgame-Perfect Guarantee and Theoretical Properties

SPPPA exactly realizes the subgame-perfect criterion: at any iteration nn and for every realized history Hn1\mathcal{H}_{n-1}, the algorithm’s actions attain the smallest possible maximal worst-case normalized suboptimality for the continuation problem. This property is captured by

minAAHn1max(f,x0) consistent with Hn1f(yNA)f12x0y2=1τn,N\min_{A \in \mathcal{A}^{\mathcal{H}_{n-1}}} \max_{(f, x_0) \text{ consistent with } \mathcal{H}_{n-1}} \frac{ f(y_N^A) - f_\star }{ \frac12 \|x_0 - y_\star\|^2 } = \frac{1}{\tau_{n, N}}

where τn,N\tau_{n, N} is the terminal progress parameter, computed from the current “seed” τn,n1=τ\tau_{n,n-1} = \tau' and propagated via the OPPA recurrence. For the initial step, this reduces to the standard minimax guarantee, and SPPPA only improves upon this at future steps as history is revealed.

Through maintaining an appropriate inductive potential, Hn=τn(ffn)12zn+1y2+12x0y2H_n = \tau_n(f_\star - f_n) - \tfrac12\|z_{n+1} - y_\star\|^2 + \tfrac12\|x_0 - y_\star\|^2, the algorithm establishes at each step a matching upper bound. For lower bounds, a convex “hard” function is constructed for each possible observed history, using a zero-chain construction and the SOCP dual certificate (ξ,w\xi, w), to guarantee that no algorithm in A\mathcal{A} can improve on the bound delivered by SPPPA for the continuation subgame (Grimmer et al., 17 Nov 2025).

4. Computational Complexity and Practical Implementation

The per-iteration computational cost of SPPPA is governed by the size of the SOCP. At step nn, the SOCP features $2n$ variables (length-nn nonnegative vectors μ,λ\mu, \lambda_\star) and a single quadratic constraint. Notably, the problem dimension dd does not impact the SOCP's computational burden, as all matrix operations can be performed using the n×nn \times n Gram matrices ZZ,ZG,GGZ^\top Z, Z^\top G, G^\top G and vectors a,ba, b, yielding total per-iteration time O(n3)O(n^3) with interior-point methods.

A limited-memory variant, which stores only the most recent kk iterates, reduces the SOCP and per-step cost to O(k3)O(k^3), maintaining independence from both dd and NN. Gram matrices and vector quantities are updated at each step in O(n2)O(n^2) time. The algorithm maintains only a short prefix of the full search history—specifically (yi,gi,zi+1,τi)(y_i, g_i, z_{i+1}, \tau_i)—and calls a standard SOCP solver with a single rotated-cone constraint of size nn.

SPPPA generalizes several classical methods. In the special case where the sequence of proximal parameters is constant (LnLL_n \equiv L), each prox-step coincides with a gradient-step on the Moreau envelope of ff with step size $1/L$. Thus, SPPPA reduces exactly to the Subgame Perfect Gradient Method (SPGM), recovering its planning subproblem and corresponding dynamic improvements over classical Nesterov’s methods.

Compared to the Optimized Proximal Point Algorithm (OPPA), SPPPA maintains the same minimax-optimality guarantee for the overall class at initialization, but, crucially, it never allows the worst-case bound to become worse and typically yields strictly improved guarantees on subclasses revealed by observed history—for example, when the locally realized convex function exhibits stronger convexity or smoother behavior than assumed by the global parameterization.

6. Summary and Significance in Convex Optimization

SPPPA introduces a paradigm by which first-order methods equipped with a prox-oracle can dynamically plan their actions to guarantee, for each possible subgame determined by cumulative oracle responses, the sharpest worst-case optimality bound theoretically attainable. This approach is enabled by the tractable SOCP-based planning subproblem, which depends only on the history length, guaranteeing scalability even in high-dimensional settings.

The algorithm establishes a stronger standard of optimality—subgame perfection—than classical minimax formulations, and it offers a new direction for the design of robust algorithms in convex optimization, especially in settings where revealed information can be leveraged for improved adaptation over time (Grimmer et al., 17 Nov 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Subgame Perfect Proximal Point Algorithm.