Papers
Topics
Authors
Recent
Search
2000 character limit reached

Extended Power Weiszfeld Method

Updated 17 January 2026
  • Extended Power Weiszfeld is a robust optimization algorithm that minimizes q-th power distances, generalizing the geometric median and weighted least squares.
  • It introduces a desingularization subgradient strategy to escape singularities, ensuring well-defined iterative updates and global convergence.
  • Empirical results, especially in portfolio optimization, show that intermediate q values balance robustness and convergence speed for enhanced performance.

The Extended Power Weiszfeld method is a de-singularity subgradient approach for solving the extended Weber location problem, formulated as the minimization of a sum of qq-th power distances between a variable point and a collection of weighted data points: f(x)=i=1nwixaiq,1q<2,xRd,f(x) = \sum_{i=1}^n w_i \|x - a_i\|^q, \qquad 1 \leq q < 2,\quad x\in\mathbb{R}^d, where wi>0w_i > 0 are positive weights, aiRda_i\in\mathbb{R}^d are data points, and the objective ff is strictly convex under non-collinearity assumptions. This formulation generalizes the geometric median (q=1q=1) and weighted least squares (q=2q=2) problems. Classical iterative solutions, such as the Weiszfeld algorithm, experience a breakdown at singularities (data points) for q<2q < 2. The Extended Power Weiszfeld method introduces systematic desingularization through a subgradient-based remedy, ensuring well-defined iterative progress and global convergence, including at data points.

1. Formulation and Singularity of the Extended Weber Problem

The extended Weber location problem seeks the unique minimizer xx_* of

f(x)=i=1nwixaiq,f(x) = \sum_{i=1}^n w_i \|x - a_i\|^q,

with 1q<21 \leq q < 2. For xaix \neq a_i for all ii, the gradient is

f(x)=i=1nqwixaiq2(xai).\nabla f(x) = \sum_{i=1}^n q w_i \|x-a_i\|^{q-2} (x - a_i).

The standard "power–Weiszfeld" update is defined as

x(k+1)=T1(x(k))=iwix(k)aiq2aiiwix(k)aiq2.x^{(k+1)} = T_1(x^{(k)}) = \frac{\sum_i w_i \|x^{(k)}-a_i\|^{q-2} a_i}{\sum_i w_i \|x^{(k)}-a_i\|^{q-2}}.

For q<2q < 2, if an iterate lands on aka_k, then xakq2\|x - a_k\|^{q-2}\to\infty, causing the update to be undefined ("breakdown"). This singularity obstructs convergence at and near data points.

2. The De-Singularity Subgradient Approach

To address the breakdown, a desingularized subgradient strategy is adopted.

Subgradient at a Data Point

For x=akx = a_k, isolate f(x)f(x) as wkxakq+Dq(x)w_k \|x - a_k\|^q + D_q(x), with Dq(x)=ikwixaiqD_q(x) = \sum_{i \neq k} w_i \|x - a_i\|^q. Define the qq-power desingularity subgradient as

Dq(ak)=ikqwiakaiq2(akai).\nabla D_q(a_k) = \sum_{i\neq k} q w_i \|a_k - a_i\|^{q-2} (a_k - a_i).

For $1 < q < 2$, the subdifferential at aka_k is the singleton {Dq(ak)}\{\nabla D_q(a_k)\}; for q=1q=1 it is the classical 1\ell_1-median subgradient set.

Escape from a Singular Point

If x(k)=akx^{(k)} = a_k and Dq(ak)>0\|\nabla D_q(a_k)\| > 0, move by

x(k+1)=akλDq(ak),x^{(k+1)} = a_k - \lambda \nabla D_q(a_k),

with λ\lambda chosen small enough to ensure strict decrease in ff. An initial λ0\lambda_0 can be set as

λ0=min{1,  1qwk1/(q1)Dq(ak)(2q)/(q1)},\lambda_0 = \min\left\{ 1,\; \frac{1}{q} w_k^{-1/(q-1)} \|\nabla D_q(a_k)\|^{-(2-q)/(q-1)} \right\},

and then backtracked (λρλ\lambda \leftarrow \rho \lambda with 0<ρ<10<\rho<1) until ff decreases.

Unified Iteration Scheme

The overall method operates as follows: x(k+1)={T1(x(k))x(k){ai} akλkDq(ak)x(k)=ak,  Dq(ak)0 akx(k)=ak,  Dq(ak)=0x^{(k+1)} = \begin{cases} T_1(x^{(k)}) & x^{(k)} \notin \{a_i\} \ a_k - \lambda_k \nabla D_q(a_k) & x^{(k)} = a_k,\; \nabla D_q(a_k) \neq 0 \ a_k & x^{(k)} = a_k,\; \nabla D_q(a_k) = 0 \end{cases} with backtracking ensuring f(x(k+1))<f(x(k))f(x^{(k+1)}) < f(x^{(k)}) at each step (Lai et al., 2024).

3. Convergence Properties

Global Convergence

Under the assumptions wi>0w_i > 0 and non-collinear {ai}\{a_i\}, the sequence {x(k)}\{x^{(k)}\} generated by this rule exhibits:

  • Monotonic decrease: f(x(k+1))<f(x(k))f(x^{(k+1)}) < f(x^{(k)}) until termination.
  • Fixed points: Only possible at the data points {ai}\{a_i\} and the unique minimizer xx_*.
  • Any non-optimal data point can be visited at most once; the escape step immediately returns the iterate to the regular region.
  • Global convergence: x(k)xx^{(k)} \to x_* from any starting point.

Superlinear Local Convergence at Singular Minima

If the unique minimizer xx_* coincides with a data point aka_k, then as xakx \to a_k,

Dq(x)=O(xak).\|\nabla D_q(x)\| = O(\|x - a_k\|).

Thus, near a singular minimizer and for $1 < q < 2$,

limkx(k+1)akx(k)ak=0,\lim_{k \to \infty} \frac{\|x^{(k+1)} - a_k\|}{\|x^{(k)} - a_k\|} = 0,

implying a superlinear convergence rate (Lai et al., 2024).

4. Algorithmic Structures

A structured summary of the procedure is as follows:

Scenario Update Step Stopping Criterion
x(k){ai}x^{(k)} \notin \{a_i\} x(k+1)=iwix(k)aiq2aiiwix(k)aiq2x^{(k+1)} = \frac{\sum_i w_i\|x^{(k)}-a_i\|^{q-2} a_i}{\sum_i w_i\|x^{(k)}-a_i\|^{q-2}} N/A
x(k)=akx^{(k)} = a_k, Dq(ak)<ε\|\nabla D_q(a_k)\|<\varepsilon x=akx_* = a_k (optimum found) Dq(ak)<ε\|\nabla D_q(a_k)\|<\varepsilon
x(k)=akx^{(k)} = a_k, Dq(ak)ε\|\nabla D_q(a_k)\|\geq\varepsilon x(k+1)=akλkDq(ak)x^{(k+1)} = a_k - \lambda_k \nabla D_q(a_k) (with backtracking on λk\lambda_k until ff decreases) N/A

Typical parameters are backtracking factor ρ=0.1\rho=0.1 and tolerance 10910^{-9}.

5. Comparisons across Exponent Choices

  • q=1q = 1: Reduces to the classical Weiszfeld algorithm ("geometric median") with the Kuhn–Weiszfeld escape step for singularities.
  • q=2q = 2: Exact minimizer in one step at the weighted Euclidean mean.
  • $1 < q < 2$: Interpolates between robustness (q=1q=1) and speed (q=2q=2), achieving faster (superlinear) convergence near singular minima, with an automatically scaled linesearch step.

A plausible implication is that selecting intermediate values q1.2q \approx 1.2–$1.6$ provides a balance between robustness to outliers and iterative efficiency.

6. Empirical Evaluation in Portfolio Optimization

In online portfolio selection:

  • At each trading day, compute the qq-median of the last mm price-relative vectors via this method and use as forecast in a reversion strategy.
  • Evaluated on NYSE(N) (d=23d=23, T6000T\approx6000) and CSI300 (Chinese index, d=47d=47, T500T\approx500), for m=5m = 5 or $10$, over q=1.1q = 1.1 to $1.9$.
  • The algorithm does not stall and escapes singularities efficiently (\approx2–3 inner backtrackings).
  • Iteration count is modest (30\leq 30), CPU time per median below $1$ ms.
  • Empirical rate of convergence decreases from λ0.35\lambda\approx 0.35 at q=1q=1 to λ0.06\lambda\approx 0.06 at q=1.9q=1.9.
  • Using qq-median forecasts in reversion portfolios, intermediate qq (1.3\approx1.3--$1.6$) often yields better cumulative returns and Sharpe ratios than both q=1q=1 and q=2q=2 (Lai et al., 2024).

7. Practical Recommendations and Significance

  • The de-singularity escape step guarantees robustness against iterate stalling and adds negligible computational overhead.
  • For online learning, machine learning, and optimization applications that generalize the median/minimum norm paradigm, intermediate values of qq are recommended for an optimal robustness–convergence trade-off.
  • Theoretical guarantees include monotonic decrease, avoidance of singularity lock-in, and superlinear local convergence at most problematic cases.

The Extended Power Weiszfeld method thus advances both theoretical analysis and practical computation for the extended Weber location problem in the non-quadratic, nonconvex setting of 1q<21 \leq q < 2 (Lai et al., 2024).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Extended Power Weiszfeld.