Elastic Tensor Factorization
- Elastic tensor factorization is a method that integrates CP/PARAFAC decomposition with elastic net regularization to enforce both sparsity and low-rank constraints.
- It employs block coordinate descent and stochastic optimization to efficiently balance data fidelity with regularization, enabling simultaneous model selection.
- Empirical evaluations demonstrate high recovery scores and improved clustering accuracy in synthetic, imaging, and clinical datasets compared to standard approaches.
Elastic tensor factorization refers to a class of tensor factorization methods that incorporate both sparsity and low-rank structure via elastic net–type regularization, enabling simultaneous model selection on both the rank and factor sparsity of tensor decompositions. This approach is notably instantiated in the context of the CANDECOMP/PARAFAC (CP) model, where the combination of and regularization on the factor matrices results in robust recovery even under noise and missing data, as well as controllable trade-offs between sparsity, rank, and reconstruction error (Shi et al., 2017).
1. The Elastic Net Regularized PARAFAC Model
The canonical model underlying elastic tensor factorization is the CP/PARAFAC decomposition of an -way tensor as
where and is the -th column. The elastic tensor factorization framework extends this model by assuming access to (possibly incomplete and noisy) observations masked by , and estimating the factor matrices by minimizing a penalized loss with elastic net regularization:
where trades off regularization and data fidelity, and tunes between sparsity () and ridge () penalties (Shi et al., 2017).
2. Bayesian Formulation and Model Selection
The elastic tensor factorization objective arises as the MAP estimator under a Gaussian likelihood for the observed entries and a factor-column prior given by a product of zero-mean Gaussian and Laplace distributions. The trace constraint is imposed across modes for identifiability, and the elastic net controls the effective tensor rank and the sparsity pattern within and across factors. As the regularization parameter increases, both the estimated tensor rank and factor matrix sparsity increase, providing a path for simultaneous model selection via regularization.
A key feature is the solution-path strategy: by computing solutions for a grid of values, one tracks the stabilization of both rank and sparsity patterns. When a plausible configuration emerges, a final sparse-constrained optimization with small regularization is performed to minimize the LS error (Shi et al., 2017).
3. Algorithms: Block Coordinate and Stochastic Optimization
The principal algorithmic approaches are block coordinate descent (BCD) and its stochastic variant. For each rank- component and each mode- factor, the BCD update involves solving
where and are derived from the current residual, mask, and Khatri–Rao products of other factors. The stochastic BCD incorporates mini-batch sampling and Adamax-style adaptive steps to scale to large tensors, and convergence to a critical point is established for the full BCD under standard assumptions. The stochastic variant is heuristic but effective in practice for large-scale data.
The per-iteration complexity is , making stochastic sampling advantageous for high-dimensional tensors (Shi et al., 2017).
4. Empirical Evaluation and Applications
Elastic tensor factorization demonstrates robust recovery and model selection in synthetic and real data contexts. Key experimental findings include:
- On synthetic tensors (, SNR 20 dB), the elastic net solution achieves average recovery score with negligible excess error versus standard CP–ALS () and LRTI ().
- In the presence of missing data (), the method with recovers the true rank $30/30$ times versus $10/30$ for LRTI.
- On COIL-20 images, clustering accuracy using the third-mode factors is $75$– (elastic net) versus $60$– (CP–ALS).
- For COPDGene clinical data analysis, the best solution yielded rank $328$, relative error $0.137$, and factor sparsity, successfully identifying salient clinical features (Shi et al., 2017).
5. Practical Implementation Aspects
Selection of in allows practitioners to trade off rank and sparsity. The regularization parameter is typically explored on a logarithmic scale. Initialization strategies include random Gaussian and leading singular vectors for each mode unfolding, though the latter may not favor sparsity. Warm starts along the regularization path improve efficiency.
The mode-specific covariance matrices can be estimated from data via Gram matrices if unknown, using the relationship
Stopping criteria include the relative error on held-out entries or stability of the active set (Shi et al., 2017).
6. Significance, Limitations, and Research Context
The elastic net strategy provides a convexified model selection mechanism on both rank and factor sparsity within the CP/PARAFAC framework, improving robustness to noise and missing data. Convergence guarantees are available for the BCD algorithm under standard conditions. The stochastic optimization approach offers scalability, with the practical caveat that convergence guarantees do not extend to the stochastic (Adamax) variant.
A plausible implication is that the solution path approach enables identification of both low-rank and sparse structure without extensive tuning or prior knowledge of the true model complexity. However, optimal recovery is not guaranteed in every scenario, and computational cost is higher than unregularized ALS due to the non-separability of the elastic net penalty and the need for path exploration. Further, the approach is defined concretely only for PARAFAC/CP factorization and does not address other tensor network formats.
7. Relation to Other Forms of Tensor Factorization
Elastic tensor factorization is distinct from harmonic factorization of elasticity tensors, which is concerned with equivariant decompositions of fourth-order harmonic tensors into lower-order covariants for the reconstruction and analysis of the elasticity tensor in continuum mechanics (Olive et al., 2016). The elastic net–based approach is independent of tensor symmetry classes and instead operates directly on arbitrary -way tensors through regularized CP decomposition (Shi et al., 2017).