Papers
Topics
Authors
Recent
Search
2000 character limit reached

Distributed Tensor Principal Component Analysis with Data Heterogeneity

Published 19 May 2024 in stat.ME, math.ST, and stat.TH | (2405.11681v3)

Abstract: As tensors become widespread in modern data analysis, Tucker low-rank Principal Component Analysis (PCA) has become essential for dimensionality reduction and structural discovery in tensor datasets. Motivated by the common scenario where large-scale tensors are distributed across diverse geographic locations, this paper investigates tensor PCA within a distributed framework where direct data pooling is impractical. We offer a comprehensive analysis of three specific scenarios in distributed Tensor PCA: a homogeneous setting in which tensors at various locations are generated from a single noise-affected model; a heterogeneous setting where tensors at different locations come from distinct models but share some principal components, aiming to improve estimation across all locations; and a targeted heterogeneous setting, designed to boost estimation accuracy at a specific location with limited samples by utilizing transferred knowledge from other sites with ample data. We introduce novel estimation methods tailored to each scenario, establish statistical guarantees, and develop distributed inference techniques to construct confidence regions. Our theoretical findings demonstrate that these distributed methods achieve sharp rates of accuracy by efficiently aggregating shared information across different tensors, while maintaining reasonable communication costs. Empirical validation through simulations and real-world data applications highlights the advantages of our approaches, particularly in managing heterogeneous tensor data.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (52)
  1. Tensors in statistics. Annual Review of Statistics and its Application 8, 345–368.
  2. Protein function prediction via graph kernels. Bioinformatics 21, i47–i56.
  3. On optimizing distributed Tucker decomposition for ssarse tensors. In Proceedings of the 2018 International Conference on Supercomputing.
  4. Communication-efficient distributed eigenspace estimation. SIAM Journal on Mathematics of Data Science 3(4), 1067–1092.
  5. Charkaborty, A. and V. M. Panaretos (2022). Testing for the rank of a covariance operator. The Annals of Statistics 50(6), 3510–3537.
  6. Semi-parametric tensor factor analysis by iteratively projected singular value decomposition. Journal of the Royal Statistical Society Series B: Statistical Methodology, To appear.
  7. Distributed estimation for principal component analysis: An enlarged eigenspace analysis. Journal of the American Statistical Association 117(540), 1775–1786.
  8. Quantile regression under memory constraint. The Annals of Statistics 47(6), 3244–3273.
  9. Chen, X. and M.-g. Xie (2014). A split-and-conquer approach for analysis of extraordinarily large data. Statistica Sinica, 1655–1684.
  10. Selecting the number of principal components: Estimation of the true rank of a noisy matrix. The Annals of Statistics, 2590–2617.
  11. Davis, C. and W. M. Kahan (1970). The rotation of eigenvectors by a perturbation. III. SIAM Journal on Numerical Analysis 7(1), 1–46.
  12. A multilinear singular value decomposition. SIAM Journal on Matrix Analysis and Applications 21(4), 1253–1278.
  13. On the best rank-1 and rank-(R1,R2,…,RN)subscript𝑅1subscript𝑅2…subscript𝑅𝑁(R_{1},R_{2},\dots,R_{N})( italic_R start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_R start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , … , italic_R start_POSTSUBSCRIPT italic_N end_POSTSUBSCRIPT ) approximation of higher-order tensors. SIAM Journal on Matrix Analysis and Applications 21(4), 1324–1342.
  14. Distributed estimation of principal eigenspaces. The Annals of Statistics 47(6), 3009–3031.
  15. A review of distributed statistical inference. Statistical Theory and Related Fields 6(2), 89–99.
  16. Communication-efficient algorithms for distributed stochastic principal component analysis. In Proceedings of the 34th International Conference on Machine Learning.
  17. Tensor analysis with n𝑛nitalic_n-mode generalized difference subspace. Expert Systems with Applications 171, 114559.
  18. Rank determination in tensor factor model. Electronic Journal of Statistics 16(1), 1726–1803.
  19. Tensor principal component analysis via sum-of-square proofs. In Proceedings of the 28th Conference on Learning Theory.
  20. Communication-efficient distributed covariance sketch, with application to distributed PCA. Journal of Machine Learning Research 22(80), 1–38.
  21. D-Tucker: Fast and memory-efficient tucker decomposition for dense tensors. In 2020 IEEE 36th International Conference on Data Engineering (ICDE).
  22. Communication-efficient distributed statistical inference. Journal of the American Statistical Association 114(526), 668–681.
  23. Multiverse recommendation: N-dimensional tensor factorization for context-aware collaborative filtering. In Proceedings of the fourth ACM conference on Recommender systems.
  24. Kolda, T. G. and B. W. Bader (2009). Tensor decompositions and applications. SIAM Review 51(3), 455–500.
  25. Communication-efficient sparse regression. Journal of Machine Learning Research 18(5), 1–30.
  26. Communication-efficient distributed optimization in networks with gradient tracking and variance reduction. Journal of Machine Learning Research 21(180), 1–51.
  27. Statistical inference in massive data sets. Applied Stochastic Models in Business and Industry 29(5), 399–409.
  28. Tucker tensor regression and neuroimaging analysis. Statistics in Biosciences 10(3), 520–545.
  29. Distributed adaptive Huber regression. Computational Statistics & Data Analysis 169, 107419.
  30. Debiased distributed learning for sparse partial linear models in high dimensions. Journal of Machine Learning Research 23(2), 1–32.
  31. TUDataset: A collection of benchmark datasets for learning with graphs.
  32. Statistical limits of spiked tensor models. Annales de l’Institut Henri Poincaré, Probabilités et Statistiques 56(1), 230–264.
  33. Blockchain-powered tensor meta-learning-driven intelligent healthcare system with IoT assistance. IEEE Transactions on Network Science and Engineering 10(5), 2503–2513.
  34. A statistical model for tensor PCA. In Advances in Neural Information Processing Systems, Volume 27.
  35. A massive data framework for M-estimators with cubic-rate. Journal of the American Statistical Association 113(524), 1698–1709.
  36. Fully scalable methods for distributed tensor factorization. IEEE Transactions on Knowledge and Data Engineering 29(1), 100–113.
  37. Tensors in Modern Statistical Learning, pp.  1–25.
  38. Communication-constrained distributed quantile regression with optimal statistical guarantees. Journal of Machine Learning Research 23(272), 1–61.
  39. Vershynin, R. (2010). Introduction to the non-asymptotic analysis of random matrices. arXiv preprint arXiv:1011.3027.
  40. Distributed inference for quantile regression processes. The Annals of Statistics 47(3), 1634–1662.
  41. Learning from binary multiway data: Probabilistic tensor decomposition and its statistical optimality. Journal of Machine Learning Research 21(154), 1–38.
  42. Tensor-view topological graph neural network. In Proceedings of the 27th International Conference on Artificial Intelligence and Statistics.
  43. Xia, D. (2021). Normal approximation and confidence region of singular subspaces. Electronic Journal of Statistics 15(2), 3798–3851.
  44. Inference for low-rank tensors – no need to debias. The Annals of Statistics 50(2), 1220–1245.
  45. The power of preconditioning in overparameterized low-rank matrix sensing. In Proceedings of the 40th International Conference on Machine Learning.
  46. Simultaneous inference for massive data: Distributed bootstrap. In Proceedings of the 37th International Conference on Machine Learning.
  47. Distributed bootstrap for simultaneous inference under high dimensionality. Journal of Machine Learning Research 23(195), 1–77.
  48. A useful variant of the Davis–Kahan theorem for statisticians. Biometrika 102(2), 315–323.
  49. Optimal sparse singular value decomposition for high-dimensional high-order data. Journal of the American Statistical Association 114(528), 1708–1725.
  50. Tensor SVD: Statistical and computational limits. IEEE Transactions on Information Theory 64(11), 7311–7338.
  51. Debiasing and distributed estimation for high-dimensional quantile regression. IEEE Transactions on Neural Networks and Learning Systems 31(7), 2569–2577.
  52. Limit results for distributed estimation of invariant subspaces in multiple networks inference and PCA. arXiv preprint arXiv:2206.04306.
Citations (3)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.