Geodesic Convexity of the Symmetric Eigenvalue Problem and Convergence of Riemannian Steepest Descent
Abstract: We study the convergence of the Riemannian steepest descent algorithm on the Grassmann manifold for minimizing the block version of the Rayleigh quotient of a symmetric matrix. Even though this problem is non-convex in the Euclidean sense and only very locally convex in the Riemannian sense, we discover a structure for this problem that is similar to geodesic strong convexity, namely, weak-strong convexity. This allows us to apply similar arguments from convex optimization when studying the convergence of the steepest descent algorithm but with initialization conditions that do not depend on the eigengap $\delta$. When $\delta>0$, we prove exponential convergence rates, while otherwise the convergence is algebraic. Additionally, we prove that this problem is geodesically convex in a neighbourhood of the global minimizer of radius $\mathcal{O}(\sqrt{\delta})$.
- Riemannian geometry of Grassmann manifolds with a view on algorithmic computation. Acta Applicandae Mathematicae, 80(2):199–220, 2004.
- Optimization Algorithms on Matrix Manifolds. Princeton University Press, Princeton, NJ, 2008.
- Riemannian perspective on matrix factorization. arXiv preprint arXiv:2102.00937, 2021.
- Distributed principal component analysis with limited communication. Advances in Neural Information Processing Systems, 34, 2021.
- A continuous-time perspective for modeling acceleration in riemannian optimization. In International Conference on Artificial Intelligence and Statistics, pages 1297–1307. PMLR, 2020.
- Momentum improves optimization on riemannian manifolds. In International Conference on Artificial Intelligence and Statistics, pages 1351–1359. PMLR, 2021.
- A Grassmann manifold handbook: Basic geometry and computational aspects. arXiv:2011.13699 [cs, math], December 2020.
- Nicolas Boumal. An introduction to optimization on smooth manifolds. To appear with Cambridge University Press, Apr 2022.
- A note on nesterov’s accelerated method in nonconvex optimization: a weak estimate sequence approach. arXiv preprint arXiv:2006.08548, 2020.
- Numerical computation of an analytic singular value decomposition of a matrix valued function. Numerische Mathematik, 60(1):1–39, 1991.
- The geometry of algorithms with orthogonality constraints. SIAM Journal on Matrix Analysis and Applications, 20(2):303–353, 1999.
- Matrix computations. JHU press, 2013.
- The noisy power method: A meta algorithm with applications. Advances in neural information processing systems, 27, 2014.
- A method of gradients for the calculation of the characteristic roots and vectors of a real symmetric matrix. Journal of Research of the National Bureau of Standards, 1951.
- N. J. Higham and S. Cheng. Modifying the inertia of matrices arising in optimization. Lin. Alg. Appl., 275–276:261–279, 1998.
- Topics in Matrix Analysis. Cambridge University Press, 1991.
- Matrix Analysis. Cambridge University Press, Cambridge ; New York, 2nd ed edition, 2012.
- Communication-efficient distributed PCA by Riemannian optimization. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 4465–4474. PMLR, 13–18 Jul 2020.
- On exact estimates of the convergence rate of the steepest ascent method in the symmetric eigenvalue problem. Linear Algebra and its Applications, 154-156:245–257, 1991.
- J. Kuczynski and H. Wozniakowski. Estimating the largest eigenvalue by the power and lanczos algorithms with a random start, 1992.
- Inequalities on the Singular Values of an Off-Diagonal Block of a Hermitian Matrix. Journal of Inequalities and Applications, 3(2):137–142, 1999.
- Landscape correspondence of empirical and population risks in the eigendecomposition problem. IEEE Transactions on Signal Processing, 70:2985–2999, 2022.
- Ross A. Lippert. Fixing two eigenvalues by a minimal perturbation. Linear Algebra and its Applications, 406:177–200, September 2005.
- Primal–dual accelerated gradient methods with small-dimensional relaxation oracle. Optimization Methods and Software, pages 1–38, 2020.
- Convergence analysis of gradient iterations for the symmetric eigenvalue problem. SIAM J. Matrix Analysis Applications, 32:443–456, 04 2011.
- Iterative minimization of the rayleigh quotient by block steepest descent iterations. Numerical Linear Algebra with Applications, 21(5):604–617, 2014.
- Estimating the largest eigenvalue of a positive definite matrix. Mathematics of Computation, 33(148):1289–1292, 1979.
- Towards a Generalized Singular Value Decomposition. SIAM Journal on Numerical Analysis, 18(3):398–405, June 1981.
- Unitarily Invariant Metrics on the Grassmann Space. SIAM Journal on Matrix Analysis and Applications, 27(2):507–531, January 2005.
- Y. Saad. Numerical Methods for Large Eigenvalue Problems. SIAM, 2nd edition edition, 2011.
- Optimization algorithms on the grassmann manifold with application to matrix eigenvalue problems. Japan Journal of Industrial and Applied Mathematics, 31:355–400, 2014.
- Yung-Chow Wong. Sectional curvatures of Grassmann manifolds. Proceedings of the National Academy of Sciences, 60(1):75–79, May 1968.
- Riemannian svrg: Fast stochastic optimization on riemannian manifolds. Advances in Neural Information Processing Systems, 29, 2016.
- First-order Methods for Geodesically Convex Optimization. arXiv:1602.06053 [cs, math, stat], February 2016.
- Towards riemannian accelerated gradient methods. arXiv preprint arXiv:1806.02812, 2018.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.