Subgame Perfect Proximal Point Algorithm
- SPPPA is a first-order method for nonsmooth convex optimization that uses a prox-oracle and history-dependent SOCP planning to achieve subgame perfection.
- The algorithm dynamically re-optimizes its continuation strategy at each iteration, ensuring improved worst-case performance beyond classical minimax guarantees.
- A limited-memory variant reduces computational cost by focusing on recent iterates, enabling efficient updates via Gram matrices and interior-point methods.
The Subgame Perfect Proximal Point Algorithm (SPPPA) is a first-order method designed for nonsmooth convex optimization utilizing a proximal operator oracle. SPPPA achieves the game-theoretic notion of subgame perfection, a property that requires not only minimax-optimality over the original problem class but also optimality on every subclass revealed through the observed information at each step. At every iteration, SPPPA dynamically re-optimizes its continuation strategy by solving a small, history-dependent second-order cone program (SOCP), which yields performance bounds that are guaranteed never to be worse and are typically strictly better than minimax-optimal guarantees for the original class (Grimmer et al., 17 Nov 2025).
1. Problem Setting and Oracle Model
The target is unconstrained convex minimization: where is closed, proper, convex, and attains its minimum at some . The algorithm is initialized at , and progress is evaluated by normalized suboptimality: The model assumes access to a “prox-oracle of accuracy one”: for any query , the oracle returns
where is a predefined sequence of positive proximal parameters.
Algorithms in this class, denoted , are deterministic and at each step query the prox-oracle, then select their next iterate within the affine span generated by past subgradients: with this selection depending solely on previously observed history .
2. The Subgame Perfect Proximal Point Algorithm: Structure and Update
SPPPA generalizes the Optimized Proximal Point Algorithm (OPPA) by dynamically re-optimizing its proof of progress at each iteration. While OPPA proceeds by a closed-form recurrence, SPPPA, at each iteration , observes the current history
and constructs matrices and vectors encoding this history:
- Select the index with minimal observed value
- Form , in
- Vectors: ,
- Auxiliary vectors:
The next step is determined by solving the following SOCP in non-negative variables :
Denote optima by , then
With , update as
3. Subgame-Perfect Guarantee and Theoretical Properties
SPPPA exactly realizes the subgame-perfect criterion: at any iteration and for every realized history , the algorithm’s actions attain the smallest possible maximal worst-case normalized suboptimality for the continuation problem. This property is captured by
where is the terminal progress parameter, computed from the current “seed” and propagated via the OPPA recurrence. For the initial step, this reduces to the standard minimax guarantee, and SPPPA only improves upon this at future steps as history is revealed.
Through maintaining an appropriate inductive potential, , the algorithm establishes at each step a matching upper bound. For lower bounds, a convex “hard” function is constructed for each possible observed history, using a zero-chain construction and the SOCP dual certificate (), to guarantee that no algorithm in can improve on the bound delivered by SPPPA for the continuation subgame (Grimmer et al., 17 Nov 2025).
4. Computational Complexity and Practical Implementation
The per-iteration computational cost of SPPPA is governed by the size of the SOCP. At step , the SOCP features $2n$ variables (length- nonnegative vectors ) and a single quadratic constraint. Notably, the problem dimension does not impact the SOCP's computational burden, as all matrix operations can be performed using the Gram matrices and vectors , yielding total per-iteration time with interior-point methods.
A limited-memory variant, which stores only the most recent iterates, reduces the SOCP and per-step cost to , maintaining independence from both and . Gram matrices and vector quantities are updated at each step in time. The algorithm maintains only a short prefix of the full search history—specifically —and calls a standard SOCP solver with a single rotated-cone constraint of size .
5. Special Cases and Connections to Related Algorithms
SPPPA generalizes several classical methods. In the special case where the sequence of proximal parameters is constant (), each prox-step coincides with a gradient-step on the Moreau envelope of with step size $1/L$. Thus, SPPPA reduces exactly to the Subgame Perfect Gradient Method (SPGM), recovering its planning subproblem and corresponding dynamic improvements over classical Nesterov’s methods.
Compared to the Optimized Proximal Point Algorithm (OPPA), SPPPA maintains the same minimax-optimality guarantee for the overall class at initialization, but, crucially, it never allows the worst-case bound to become worse and typically yields strictly improved guarantees on subclasses revealed by observed history—for example, when the locally realized convex function exhibits stronger convexity or smoother behavior than assumed by the global parameterization.
6. Summary and Significance in Convex Optimization
SPPPA introduces a paradigm by which first-order methods equipped with a prox-oracle can dynamically plan their actions to guarantee, for each possible subgame determined by cumulative oracle responses, the sharpest worst-case optimality bound theoretically attainable. This approach is enabled by the tractable SOCP-based planning subproblem, which depends only on the history length, guaranteeing scalability even in high-dimensional settings.
The algorithm establishes a stronger standard of optimality—subgame perfection—than classical minimax formulations, and it offers a new direction for the design of robust algorithms in convex optimization, especially in settings where revealed information can be leveraged for improved adaptation over time (Grimmer et al., 17 Nov 2025).