Closed-Form Geodesics in Matrix Manifolds
- Closed-form geodesics are explicit analytical solutions for distance‐minimizing curves on Riemannian manifolds, particularly in matrix manifolds like the Stiefel, Grassmann, and flag manifolds.
- They employ a two-parameter family of Riemannian metrics and offer both block-exponential and reduced formulations to efficiently compute geodesic paths, exponential maps, and logarithm maps.
- The framework integrates Fréchet derivatives with trust-region minimization, yielding robust convergence in optimization and statistical applications on low-dimensional settings.
Closed-form geodesics are explicit analytical solutions for geodesics—locally distance-minimizing curves—on Riemannian manifolds, with particular focus on matrix manifolds such as the Stiefel, Grassmann, and flag manifolds. These formulas enable efficient computation of geodesics, exponential and logarithm maps for a two-parameter family of Riemannian metrics, crucial for algorithms in optimization and statistics on these spaces. The development of closed-form geodesics, together with Fréchet derivatives and trust-region minimization, forms a unified framework for calculating Riemannian logarithms and geodesic distances even when closed-form logarithm maps are unavailable (Nguyen, 2021).
1. Stiefel Manifold and Metric Family
The Stiefel manifold consists of orthonormal -frames in (real or complex). The tangent space at comprises matrices such that is antisymmetric. The Riemannian metric is parameterized by two positive scalars : $g_Y(\omega, \xi) = \Tr_R[ \omega^t (\alpha_0 I_n + (\alpha_1 - \alpha_0) Y Y^t) \xi ]$ The metric interpolates between the embedded metric () and canonical metric (). The associated raising operator has the form
This parameterization enables closed-form geodesic solutions that work efficiently for low-rank settings () (Nguyen, 2021).
2. Geodesic ODE and Closed-Form Solutions
The geodesic satisfies the second-order ODE: Initial conditions: , , and the decomposition: With and matrix
the geodesic admits two equivalent closed forms:
- Block-exponential (dimension ):
- Reduced (dimension $2p$):
with
Both representations maintain computational complexity at per matrix exponential, avoiding dependence on the ambient dimension when (Nguyen, 2021).
3. Fréchet Derivatives and Gradient Computation
For the Riemannian logarithm, the problem reduces to finding such that , minimizing
The gradient is obtained via Fréchet derivatives: This yields gradients with computational cost comparable to evaluating one exponential plus its Fréchet derivative (approximately three times the cost of a single exponential). The trace–Fréchet adjoint identity is applied: $\Tr[ C\,\mathcal L_f(A, E)\,D ] = \Tr[ \mathcal L_f(A, D C)\,E ]$ The gradient splits into blocks: with , enabling efficient exact gradient evaluation (Nguyen, 2021).
4. Trust-Region Minimization and the Riemannian Logarithm
The Riemannian logarithm is computed by minimizing over antisymmetric and arbitrary . Standard interior-point or trust-region solvers (e.g., SciPy’s “trust-krylov”) are employed, leveraging the closed-form gradient. When is within the injectivity radius of , the Hessian of is positive definite and convergence to the unique minimizing logarithm is rapid. For moderately large distances, the algorithm often still recovers nearly minimal “shooting” geodesics.
This approach extends to quotient manifolds:
- On , for a subgroup (e.g., Grassmann or flag), restrict to the horizontal subspace and adapt cost functions to measure ambient-coordinate distance.
- The trust-region and Fréchet framework applies identically for these cases (Nguyen, 2021).
5. Grassmann and Flag Manifolds: Trigonometric Formulas
For the Grassmann manifold , the horizontal space is . For a purely horizontal geodesic (): $\Exp^{Gr}_Y(\eta) = \Exp^{St}_Y(\eta) = Y\,\cos(S^{1/2}) + \eta\,S^{-1/2} \sin(S^{1/2}), \quad S = \eta^t\eta$ Given , compute the SVD: The unique minimizing logarithm for all : $\Log_Y^{Gr}(Z) = (Z - Y Y^t Z)\; V\; (I - \Sigma^2)^{-1/2} \arccos(\Sigma) U^t$ The geodesic distance is
$d_{Gr}(Y, Z) = \sqrt{\alpha_0\,\Tr[\arccos^2(\Sigma)]}$
These are analogous to the , formulas on the sphere (Nguyen, 2021).
6. Algorithmic Complexity, Stability, and Applications
Each trust-region iteration requires one or two matrix exponentials (size at most ) plus Fréchet derivatives; cost per exponential is . For low-rank problems (), all computations occur in small dimension (). Fréchet-derivative routines (e.g., SciPy's expm_frechet) compute both the exponential and its Fréchet derivative at threefold the cost of the exponential alone.
The approach avoids integration along curves or multi-step shooting, instead solving a small-dimensional nonlinear least-squares problem. Numerical experiments on Stiefel and flag manifolds demonstrate robust convergence for distances up to and beyond half the injectivity radius. The Hessian remains well-conditioned for points near one another, deteriorating only as one approaches the cut locus.
One immediate application is computation of the Riemannian center of mass (Karcher mean): minimizing by gradient descent, where each step requires logarithm computations—now feasible in low-dimension settings. The methods are conceptually and algorithmically unified, supporting practical statistical and optimization tasks on matrix manifolds (Nguyen, 2021).