Papers
Topics
Authors
Recent
Search
2000 character limit reached

Riemannian Autoencoders Overview

Updated 27 January 2026
  • Riemannian autoencoders are generative frameworks that model the latent space as a differentiable manifold using decoder-induced pull-back metrics to capture intrinsic geometry.
  • They integrate geometry-aware priors, reparameterization via exponential maps, and tailored regularization to improve ELBO and produce higher-quality interpolations and samples.
  • These models extend to complex, mixed-curvature, and Lie group latent spaces, offering robust representation learning and practical benefits for high-dimensional data.

Riemannian autoencoders are generative learning frameworks in which the latent space is endowed with a Riemannian manifold structure, rather than the standard Euclidean geometry. This geometric formalism leverages decoder-induced or data-driven metrics to respect intrinsic distances, curvature, and topology of the data distribution, yielding enhanced sample quality, more plausible interpolations, improved representation learning, and the ability to model nontrivial latent geometries, including constant or mixed curvature, complex (Kähler), and data-induced metrics. Riemannian autoencoders comprise both variational and non-variational approaches, with core methodologies built upon pull-back metrics (often from the decoder), intrinsic manifold measures, manifold-informed priors and posteriors, and geometry-aware regularization and losses.

1. Mathematical Foundations: Pull-Back Metrics and Manifold Structure

The essential principle of Riemannian autoencoders is to reinterpret the latent space Z of an autoencoder as a differentiable manifold (𝓜, g), where g is a Riemannian metric tensor. This is most classically constructed via the pull-back mechanism: for a differentiable decoder map μₜₕₑₜₐ : 𝓜→ℝᵐ (as in a VAE), the metric at z∈𝓜 is

G(z)=Jμ(z)Jμ(z)+Jσ(z)Jσ(z)G(z) = J_{μ}(z)^\top J_{μ}(z) + J_{σ}(z)^\top J_{σ}(z)

where J_{μ} and J_{σ} are Jacobians of the decoder mean and variance, respectively (Kalatzis et al., 2020). This metric equips the tangent space T_z𝓜 with inner product ⟨v, w⟩_z = v\top G(z) w. For some models, only the mean map is used; for others, decoder uncertainty or learned metric networks supplement or replace the explicit Jacobian terms (Chadebec et al., 2020).

The Riemannian metric enables definition of geodesic curves γ : [0,1]→𝓜 by minimizing the energy functional:

E(γ)=1201γ˙(t)G(γ(t))γ˙(t)dt,E(\gamma) = \frac{1}{2} \int_0^1 \dot{\gamma}(t)^\top G(\gamma(t)) \dot{\gamma}(t) \, dt,

resulting in the Riemannian distance d(z, z′) as the minimal energy path length.

These constructions extend to specialized settings:

2. Riemannian Priors, Posteriors, and Reparameterization Strategies

To align prior and posterior models with intrinsic geometry, Riemannian autoencoders replace the standard flat Gaussian prior with manifold-adapted distributions:

  • Brownian motion/heat kernel priors: On (𝓜, g), the prior is given by the transition density of Riemannian Brownian motion:

pt(zz0)=Zt1Kt(z0,z),Kt(z0,z)(2πt)n/2H0(z,z0)exp(d(z,z0)22t)p_t(z | z_0) = Z_t^{-1} K_t(z_0, z), \qquad K_t(z_0, z) \approx (2\pi t)^{-n/2} H_0(z, z_0) \exp\left(-\frac{d(z, z_0)^2}{2t}\right)

where H0H_0 is the volume form ratio. This recovers standard Gaussian when 𝓜 = ℝn (Kalatzis et al., 2020).

  • Wrapped Gaussians: For homogeneous manifolds (e.g. spheres, hyperboloids), area-preserving diffeomorphisms push Euclidean Gaussians through the exponential (or Lambert/gyro) maps, preserving normalization and simple sampling (Galaz-Garcia et al., 2022, Skopek et al., 2019).
  • Product manifolds: Posteriors and priors are constructed as products of wrapped normal distributions over each factor, with full reparameterization via tangent-normal sampling, parallel transport, and exponential maps (Skopek et al., 2019).
  • Hamiltonian proposals: For samplers requiring manifold volume correction, the Riemannian Hamiltonian approach defines the auxiliary momentum in T_z𝓜, evolving (z,ρ)(z, \rho) under Hamiltonian flow specified by the metric tensor (Chadebec et al., 2020).

Reparameterization on the manifold employs either (i) the exponential map applied to Gaussian tangent noise, or (ii) approximate shooting methods (projected small steps, Jacobian pseudo-inverse back-projection) (Kalatzis et al., 2020).

3. Variational Objectives and Geometry-Aware Regularization

The evidence lower bound (ELBO) for Riemannian autoencoders incorporates densities, volumes, and KL divergences under the manifold measure. When both prior and posterior are written as heat kernels or wrapped Gaussians on 𝓜, normalization factors cancel, and KL is tractably estimated by Monte Carlo (Kalatzis et al., 2020, Galaz-Garcia et al., 2022). The Riemannian structure introduces additional terms:

  • Volume corrections: Sampling and loss terms include √det G(z) in density calculations, and geometry-aware sampling exploits the Riemannian measure (Gracyk, 19 Nov 2025, Chadebec et al., 2022).
  • Regularization: Flatness (low-bending) and isometry penalties encourage mappings that are locally distance-preserving and extrinsically as flat as possible (Braunsmann et al., 2021, Braunsmann et al., 2022). For example, loss terms penalize deviations of latent-space distances from manifold geodesics and second-difference bending vectors.
  • Decoder geometry: Additional auxiliary losses encourage the latent embedding to concentrate near the true data manifold, prevent excessive curvature, and match geometric regularity via curvature or Ricci terms (Gracyk, 19 Nov 2025, Sun et al., 2024).
  • Adversarial/geometric membership: For constant-curvature spaces, adversarial objectives force the aggregated posterior to match manifold-valued priors, and explicit constraints penalize deviation from the embedding manifold (Grattarola et al., 2018).

4. Computational Schemes: Geodesic Solvers, Interpolations, and Sampling

Computation of geodesics and interpolations in Riemannian autoencoders is central to their improved generative and representation properties. Techniques span:

5. Model Architectures, Implementation, and Empirical Findings

Architectures for Riemannian autoencoders parallel classical autoencoders but with essential geometric augmentations:

  • Encoders output manifold-structured parameters, e.g., tangent-space coordinates, means/covariances for wrapped distributions, Lie algebra elements (for group-based models) (Gong et al., 2019).
  • Decoders are differentiable immersions (usually MLPs or CNNs) whose Jacobian determines the pull-back metric. In explicit manifold models (sphere/hyperboloid), decoding may map from the manifold to data via little more than an affine or kernel network (Skopek et al., 2019, Galaz-Garcia et al., 2022).
  • Projection/denoising heads are used to approximate the manifold by an implicit projection, enabling robust implementation of Riemannian calculus even when the true manifold is not explicitly parameterized (Hartwig et al., 10 Oct 2025).

Empirical evidence demonstrates that geometry-aware priors and interpolations:

6. Extensions: Nontrivial Latent Geometries and Theoretical Guarantees

The Riemannian autoencoder framework generalizes to accommodate:

  • Complex/Kähler latent spaces: Decoders with complex-valued outputs are equipped with information-theoretic Kähler metrics, with regularization, volume sampling, and analytic curvature (Gracyk, 19 Nov 2025).
  • Mixed-curvature and Lie group latent spaces: Latent variables reside in direct products of constant curvature spaces or Lie groups, with modular reparameterization, parallel transport, and geometric regularization (Skopek et al., 2019, Gong et al., 2019).
  • Score-based/normalizing-flow models: Pullback metrics and global coordinate charts are constructed via flows or score maps, with provable reconstruction error and dimension-recovery theorems (Diepeveen et al., 2024, Diepeveen et al., 26 Jan 2026).
  • Empirical convergence and approximation guarantees: Low-distortion and low-bending regularizers are shown to Γ-converge to corresponding geometric energies, ensuring that learned latents approximate isometric and extrinsically flat embeddings (Braunsmann et al., 2022).

7. Limitations and Open Problems

While Riemannian autoencoders offer clear advantages, key limitations include:

  • Computational cost: Geodesic computation and metric inversion can pose scaling issues for high-dimensional latent spaces.
  • Approximation error: Short-time heat kernel expansions and area-preserving wraps may become inaccurate on highly curved manifolds or when latent geometry does not align with data distribution (Kalatzis et al., 2020).
  • Dependence on accurate volume and curvature estimation: Discrepancies between modeled and true geometry affect volume sampling and interpolation.
  • Requirement of ground-truth manifold distances: Some methods depend on precomputed or high-quality diffusion or graph-based distances (Sun et al., 2024).
  • Data requirements: Certain approaches (e.g., denoising projections, manifold-aware losses) require sampling or knowledge of small neighborhoods or local averages (Braunsmann et al., 2021, Braunsmann et al., 2022).

Open directions include scalable geodesic solvers for high-dimensional manifolds, joint estimation of latent curvature and dimension, geometry-aware flows for multi-modal or highly anisotropic data, and theoretical bounds on reconstruction and generalization for geometry-regularized autoencoders.


References:

Topic to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Riemannian Autoencoders.