- The paper introduces NCAA, a flexible model that relaxes the strict convexity constraint via an adjustable margin, achieving improved interpretable archetype recovery with minimal error.
- It develops a two-block coordinate descent algorithm with Nesterov acceleration to efficiently optimize matrices under generalized simplex constraints.
- Empirical results on synthetic and hyperspectral datasets demonstrate NCAA’s superior performance in mean spectral angle error compared to methods like MinVolNMF and SNPA.
Near-Convex Archetypal Analysis: Synthesis, Algorithms, and Empirical Assessment
Motivation and Context
Traditional Nonnegative Matrix Factorization (NMF) offers a computationally tractable and interpretable framework for dimensionality reduction in nonnegative data matrices, with wide deployment in areas such as hyperspectral unmixing. However, NMF's interpretability is limited because basis elements are not guaranteed to reflect actual data points; archetypal analysis (AA) and convex NMF address this by restricting basis vectors to reside in the convex hull of data points. Despite enhanced interpretability, this constraint often results in increased approximation error, especially in highly mixed or noisy settings. Prior relaxations—such as introducing convexity margins or trade-offs between reconstruction error and archetype proximity—have not fully reconciled the interpretability-versus-error trade-off, nor have they supported scalable inference with explicit coefficient matrices linking archetypes to data.
The Near-Convex Archetypal Analysis (NCAA) model proposed in this paper introduces a relaxation of the convexity constraint via an adjustable margin ϵ, yielding improved estimation accuracy while preserving the interpretability of AA.
The NCAA objective constrains archetypes to near-convex combinations (NCCs) of selected data points, allowing basis vectors to lie outside the strict convex hull via negative weights bounded by −ϵ. The formulation is as follows: for a data matrix X∈Rm×n, subset matrix Y∈Rm×d (collected via SNPA or hierarchical clustering), factorization rank r, and scalar ϵ, NCAA seeks matrices A∈Rd×r and H∈Rr×n to minimize ∥X−YAH∥F2, subject to A(k,l)≥−ϵ and ∑k=1dA(k,l)=1, with H columns in the unit simplex.
The geometric implications are formalized: as ϵ increases, the archetypes W=YA are pushed further from the convex hull of Y, parameterizing the trade-off between interpretability and fidelity. The purity level p quantifies the separation between archetypes and data points, guiding ϵ selection and indicating when exact recovery (separable NMF) is possible.
Optimization Strategy
NCAA is solved via a two-block coordinate descent (BCD), alternating projected gradient steps for A and H. Projections onto the generalized simplex (for A with negative margins, and H onto the standard simplex) employ fast projected gradient methods with Nesterov acceleration and adaptive line search for step size.
Algorithmic tuning of ϵ is achieved by initializing at a small value and doubling until the reduction in reconstruction error stalls, ensuring data-adaptive flexibility. The authors also propose column-wise fine-tuning of ϵl for each archetype, moving archetypes toward the convex hull as long as the increase in approximation error remains marginal.
The computational complexity of NCAA is linear in input dimensions per iteration (O(mnr)), provided d is chosen as a small multiple of r, making the method viable for large-scale applications.
Empirical Results
Synthetic Data Evaluation
NCAA is benchmarked against state-of-the-art minimum-volume NMF (MinVolNMF with logdet penalty and simplex constraints) and SNPA on synthetic datasets parametrized by purity p, rank r, and noise υ. The performance metric is Mean Removed Spectral Angle (MRSA), with lower values signifying better recovery. NCAA attains substantially lower MRSA than MinVolNMF and SNPA across most configurations, including low purity, high rank, and high noise scenarios. Specifically, SNPA is only competitive in ideal separable settings (p=1), confirming NCAA's robustness under practical, non-ideal distributions.
Hyperspectral Unmixing
On a real hyperspectral image (HYDICE Urban), NCAA (with hierarchical clustering for Y and fine-tuned ϵ) recovers spectral signatures closely matching ground truth endmembers. The abundance maps produced align with physical expectations, and NCAA outperforms MinVolNMF in MRSA (5.56 vs 5.73) and qualitative recovery of material mixtures. This demonstrates NCAA's practical applicability and interpretability in real-world geometric unmixing.
Implications and Future Directions
NCAA's relaxation of AA's convexity constraint formalizes the balance between interpretability and estimation error, enabling practitioners to tune fidelity via ϵ and supporting scalable inference with explicit archetype-to-data mappings. The adaptive tuning of ϵ—both globally and per-archetype—addresses variability in data purity and ensures robustness to non-separability.
The theoretical implications extend to identifiability: the proximity to minimum-volume NMF suggests that NCAA can inherit theoretical guarantees, such as uniqueness under noisy or mixed scenarios. Future work includes analysis of NCAA's uniqueness properties, extension to models with learnable Y, incorporation of row-sparsity constraints for automatic selection of archetypes, and alternate regularization schemes in the objective function.
Conclusion
The NCAA framework advances NMF-based dimensionality reduction by unifying interpretable archetype construction with low reconstruction error. Its geometric and algorithmic innovations grant practitioners both robustness and clarity in factorization, as confirmed by strong empirical results in synthetic and real data scenarios. NCAA lays the groundwork for further developments in interpretable matrix factorization, adaptive convexity relaxation, and identifiability theory.