Papers
Topics
Authors
Recent
Search
2000 character limit reached

Wasserstein Archetypal Analysis

Published 25 Oct 2022 in stat.ML, cs.LG, math.OC, and math.PR | (2210.14298v1)

Abstract: Archetypal analysis is an unsupervised machine learning method that summarizes data using a convex polytope. In its original formulation, for fixed k, the method finds a convex polytope with k vertices, called archetype points, such that the polytope is contained in the convex hull of the data and the mean squared Euclidean distance between the data and the polytope is minimal. In the present work, we consider an alternative formulation of archetypal analysis based on the Wasserstein metric, which we call Wasserstein archetypal analysis (WAA). In one dimension, there exists a unique solution of WAA and, in two dimensions, we prove existence of a solution, as long as the data distribution is absolutely continuous with respect to Lebesgue measure. We discuss obstacles to extending our result to higher dimensions and general data distributions. We then introduce an appropriate regularization of the problem, via a Renyi entropy, which allows us to obtain existence of solutions of the regularized problem for general data distributions, in arbitrary dimensions. We prove a consistency result for the regularized problem, ensuring that if the data are iid samples from a probability measure, then as the number of samples is increased, a subsequence of the archetype points converges to the archetype points for the limiting data distribution, almost surely. Finally, we develop and implement a gradient-based computational approach for the two-dimensional problem, based on the semi-discrete formulation of the Wasserstein metric. Our analysis is supported by detailed computational experiments.

Summary

  • The paper introduces a novel framework that replaces the Euclidean metric with the 2-Wasserstein distance to robustly summarize data via convex polytopes.
  • It establishes theoretical guarantees including existence, uniqueness, and statistical consistency while employing Rényi entropy for regularization.
  • Empirical evaluations demonstrate the algorithm's effectiveness on both synthetic and real datasets, highlighting its interpretability and robustness to outliers.

Wasserstein Archetypal Analysis: Theory, Algorithms, and Empirical Evaluation

Motivation and Problem Formulation

Archetypal Analysis (AA) is a classical unsupervised learning paradigm aimed at summarizing multivariate data via convex polytopes, with the polytope vertices interpreted as "archetypes"—exemplars of extreme points of the dataset. Historically, AA minimizes the average squared Euclidean distance from data points to their projections onto a convex hull contained within the convex hull of the dataset, yielding archetypes that explain the general characteristics of the data.

However, AA suffers from fundamental issues when the data-generating distribution is unbounded or contains outliers, due to the quadratic loss. To address these deficiencies, the paper introduces Wasserstein Archetypal Analysis (WAA), replacing the Euclidean metric with the 2-Wasserstein metric, thus leveraging optimal transport to robustly fit polytopal distributions to empirical or continuous data. For fixed kk, WAA seeks the convex kk-gon (or polytope in higher dimensions) whose uniform measure is closest to the data distribution μ\mu in Wasserstein distance.

Theoretical Results

The paper systematically analyzes the existence, uniqueness, and statistical consistency of the WAA minimization problem. Key results include:

  • Existence and Characterization in Low Dimensions: For d=1d=1, an explicit closed-form solution exists for WAA, utilizing the optimal transport map between μ\mu and a uniform interval, parameterized by the mean and a key moment (see Figure 1). Figure 1

    Figure 1: Illustration of limit points in the space of uniform measures on triangles, showing that narrow limits may not correspond to uniform distributions.

  • For d=2d=2, the authors prove existence of solutions under the assumption that μ\mu is absolutely continuous with respect to Lebesgue measure. The intricacies of the closure of the constraint set under the narrow topology are elucidated, including the nontrivial behavior of degenerate limits of polytope sequences.
  • Regularization via Rényi Entropy: Recognizing that minimizing sequences may collapse to lower-dimensional supports in higher dimensions or for singular measures, the paper regularizes the loss by adding a Rényi entropy term. This ensures existence and compactness of solutions for arbitrary μP2(Rd)\mu \in \mathcal{P}_2(\mathbb{R}^d) and all dd, preventing degeneracy.
  • Consistency and Empirical Convergence: The statistical consistency of WAA is established: as empirical measures μn\mu_n converge to μ\mu, the associated WAA archetypal polytopes converge (up to subsequence) to those of the limiting distribution, almost surely. The convergence rate in the objective value is quantified under moment assumptions.

These results are technically rigorous, leveraging measure-theoretic properties, optimal transport theory, and variational analysis.

Algorithmic Methods

The paper proposes a gradient-based alternating minimization algorithm for the 2D case with empirical measures, exploiting the semi-discrete optimal transport method:

  • The dual formulation of the 2-Wasserstein metric is utilized, where for discrete μ\mu and uniform polytope ν\nu, the optimal transport is computed via weighted Voronoi polygons (power diagrams), allowing direct computation and efficient optimization. Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2

Figure 2: Snapshots of the gradient-based algorithm evolving the triangle solution over iterations for uniform disk data.

  • The alternating method maximizes the dual variable over the Voronoi tessellation and descends in the space of polytope vertex positions, leveraging explicit formulae for shape derivatives of integrals over polytopes (Proposition on shape derivatives). Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3: Solutions for varying kk show regular polygons emerge as optimal archetypes for uniform disk data.

Numerical experiments demonstrate the robustness and versatility of the approach on both synthetic (uniform disk, Gaussian) and real datasets (COVID-19 positivity rates).

Empirical Analysis and Observations

  • Regular Polytopes as Archetypes: For isotropic distributions (disk, normal), the optimal WAA polytopes are empirically regular (equilateral triangles, squares), regardless of initialization. This aligns with theoretical expectations for symmetry-constrained minimization. Figure 4

Figure 4

Figure 4: Archetypal polygons for Gaussian data, showing interior solutions and dimensional consistency for k=3,4k=3,4.

  • Effects of Regularization: Increasing the Rényi entropy penalty ε\varepsilon systematically enlarges the polytope without notably distorting its shape, as verified quantitatively by the ratio of longest to shortest side (Figure 5). Figure 5

    Figure 5: Ratio of longest to shortest side decreases with ε\varepsilon, illustrating regularization effect.

  • Non-Convexity and Multiple Local Minima: The WAA landscape is numerically non-convex, with multiple stationary points evidenced in the triangle example (Figure 6), highlighting the importance of initialization and global optimization strategies. Figure 6

Figure 6

Figure 6: Energy landscape for triangle-to-triangle Wasserstein minimization exhibits local minima, indicating non-convexity.

  • Application to COVID-19 Data: WAA archetypes provide interpretable summarization of US states' pandemic trajectories in principal component space, outperforming kk-means in robustness to outliers and yielding easily interpretable exemplars.

Implications and Future Directions

Practically, WAA provides a robust alternative to classical AA for summarizing datasets with heavy tails or outliers. Theoretically, the work exposes important open questions including:

  • Extending existence proofs for ε=0\varepsilon=0 to higher dimensions and singular measures
  • Establishing conditions for uniqueness up to invariance groups
  • Generalizing to pp-Wasserstein metrics (p2p \neq 2) and other divergences
  • Developing efficient algorithms for d>2d > 2, possibly through entropic regularization or back-and-forth schemes

The analysis and methodology support WAA as a powerful, flexible tool for unsupervised summarization of multivariate distributions, particularly in settings where classical AA fails.

Conclusion

The paper rigorously formulates and analyzes Wasserstein Archetypal Analysis, providing theoretical existence and statistical consistency results, introducing a Rényi entropy regularization for generality, and developing a gradient-based semi-discrete algorithm. Empirical studies support the theoretical claims and illustrate WAA's advantages in robustness and interpretability. The implications span both methodological advances in unsupervised learning with optimal transport and practical improvements for archetype-based data summarization, paving the way for further algorithmic and theoretical development in high-dimensional and non-Euclidean settings.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.