Papers
Topics
Authors
Recent
Search
2000 character limit reached

SE(3)-Equivariant Generative Models

Updated 30 January 2026
  • SE(3)-Equivariant Generative Modelling is a framework where neural networks strictly adhere to 3D rotation and translation invariance.
  • Techniques include energy-based models, normalizing flows, and diffusion methods that leverage group theory and tensor representations for bi-equivariance.
  • Applications span robust robotic manipulation, molecular conformation, and protein design, achieving improved data efficiency and faster inference.

SE(3)-equivariant generative modelling refers to probabilistic or score-based neural generation paradigms that enforce exact equivariance to the full group of three-dimensional rotations and translations (SE(3)), typically for domains—robotics, molecules, proteins—where rigid-body invariances are physically fundamental. By design, these models guarantee that input transformations propagate predictably to output sample distributions, yielding data-efficient, robust, and generalizable generation. Techniques include energy-based models (EBMs), normalizing flows, flow matching, denoising diffusion, and their rectified and motif-augmented extensions. This article surveys foundational theories, SE(3)-equivariant architectures, group-theoretic loss formulations, representative applications in robotics/chemistry/biology, empirical metrics, and future trajectories.

1. Mathematical Foundations: SE(3) Equivariance

The special Euclidean group SE(3) comprises pairs (R, v), with RSO(3)R \in SO(3) (rotations) and vR3v \in \mathbb{R}^3 (translations), acting on points xR3x \in \mathbb{R}^3 via Tx=Rx+vT \cdot x = R x + v. Equivariance requires that model outputs transform consistently under input group actions: for any gSE(3)g \in SE(3), f(gx)=gf(x)f(g \cdot x) = g \cdot f(x).

Neural architectures achieve this via:

  • Irreducible SO(3) representations: Features decompose into ll-types, transformed via Wigner Dl(R)D_l(R) matrices (Ryu et al., 2022).
  • Complete local frames: Orthonormal bases constructed from graph neighbors (by cross products) represent atom-centric local coordinates, minimizing overhead compared to global spherical harmonics (Du et al., 2021).
  • Multiscale descriptor fields and message-passing GNNs: Inputs are decorated with scalar, vector, and higher-rank tensor features—each projected or aggregated using frame or spherical harmonic machinery for equivariance.

Formally, generative models are equivariant if the conditional density or score p(TX,Y)p(T|X,Y) satisfies:

p(STSX,Y)=p(TX,Y)=p(TSX,S1Y)p(S T | S \cdot X, Y) = p(T | X, Y) = p(T S | X, S^{-1} \cdot Y)

thus commuting with both input and output actions (bi-equivariance).

2. SE(3)-Equivariant Generative Model Types

Core model classes include:

  • Energy-Based Models: Assign an energy function E(TX,Y)E(T|X,Y) over SE(3) placements, whose Boltzmann distribution exp[E]/Z[-E]/Z is strictly SE(3)-equivariant if EE is bi-equivariant (Ryu et al., 2022).
  • Flow Models (Normalizing, Flow-Matching): Learn invertible f:R3nR3nf: \mathbb{R}^{3n} \to \mathbb{R}^{3n} such that f(gx)=gf(x)f(g \cdot x) = g \cdot f(x), enabling tractable sampling/density estimation (Midgley et al., 2023, Poletukhin et al., 23 Jan 2026, Huguet et al., 2024).
  • Denoising Diffusion and Score Models: Forward SDEs propagate samples by left-invariant Brownian motion on SE(3), while score networks predict glogpt(g)\nabla_{g} \log p_t(g); sampling integrates an annealed Langevin SDE or reverse ODE (Ryu et al., 2023).
  • Rectified and Geodesic-Consistent Flows: Trajectory-level policies on SE(3) learned by enforcing ODE geodesicity, yielding fast one-step inference (Wang et al., 20 Sep 2025).

Several models hybridize continuous flows for rotations/translations with discrete flows over fragment or motif vocabularies (e.g., rigid subgraphs in chemicals) (Poletukhin et al., 23 Jan 2026).

3. Network Architectures and Equivariance Enforcement

Representative SE(3)-equivariant architectures include:

  • Tensor Field Networks (TFN): Convolutional layers with kernels expanded in spherical harmonics and Clebsch–Gordan tensor products, enforcing rotation equivariance (Ryu et al., 2022).
  • SE(3)-Transformer & Equiformer: Attention-based layers acting on type-0 scalars and higher-type descriptors with explicit Wigner matrix propagation (Ryu et al., 2022, Ryu et al., 2023).
  • Complete Local Frames GNNs: Lightweight, per-node coordinate system (via cross products), enabling efficient, strictly equivariant message-passing (Du et al., 2021).
  • IPA (Invariant Point Attention): Pointwise attention/updates with SO(3)-equivariant handling for protein/fragment backbones (Poletukhin et al., 23 Jan 2026, Huguet et al., 2024).
  • Augmented Coupling Flows: Coordinate splits in equivariant reference frames, with element-wise splines or affine couplings in invariant bases (Midgley et al., 2023).

Network outputs are often split into translation and rotation channels, with data encoded as point clouds, frames, or fragment sets; feature fusion, auxiliary alignment losses, and motif-level symmetry handling further guarantee equivariance in complex contexts.

4. Training Objectives and Inference Procedures

Training leverages group-theoretic loss functions:

  • Maximum Likelihood (Contrastive Divergence): Gradients of log-probabilities, balanced by EBM energies on positive/negative SE(3) samples via MCMC (Ryu et al., 2022).
  • Score Matching: Denoising objectives comparing model-predicted and analytic scores of transition kernels (Gaussian for R3\mathbb{R}^3 and IGSO(3)\mathrm{IG}_{SO(3)} for rotations) (Ryu et al., 2023).
  • Flow-Matching Regression: Time-indexed regression on ODE-velocity fields interpolating source and target frames; geodesic alignment ensures loss is invariant to global/motif-level rotations (Poletukhin et al., 23 Jan 2026, Huguet et al., 2024).
  • Surrogate ELBOs: Variational bounds with noisy queries stabilize early training of EBM/descriptors (Ryu et al., 2022).

Inference and sampling employ:

  • MCMC, Langevin, and Metropolis–Hastings: Markov kernels tailored for SE(3) (orientation proposals from isotropic SO(3) normal distributions, translation by Gaussians) (Ryu et al., 2022).
  • Annealed Langevin Dynamics: Euler–Maruyama integration in SE(3)’s Lie algebra; pose update via exponentials of predicted score vectors (Ryu et al., 2023).
  • ODE Integration for Flows: Time-resolved integration of learned twist fields; single-step or adaptive schemes yield rapid trajectory prediction (Wang et al., 20 Sep 2025).
  • Discrete flows for motif classes: CTMC denoising bridges continuous generative frames with discrete motif assignment (Poletukhin et al., 23 Jan 2026).

5. Practical Applications and Empirical Evaluation

Applications span:

  • Visual Robotic Manipulation: End-to-end 6-DoF policy from point clouds; sample-efficient learning (5–10 demonstrations); generalization across unseen poses, objects, distractors; robust motion planning under SE(3) (Ryu et al., 2022, Ryu et al., 2023, Wang et al., 20 Sep 2025).
  • 3D Molecule Generation: Rigid motif-based flow matching admits compression (\sim3.5×\times fewer tokens), rapid generation (100 steps vs. \sim1000 for diffusion), higher atom/molecule stability on benchmarks (QM9, GEOM-Drugs, QMugs) (Poletukhin et al., 23 Jan 2026).
  • Protein Structure Generation: Sequence-augmented SE(3) flows generate novel backbones, with state-of-the-art designability, diversity, and novelty (FoldFlow-2), via multimodal fusion trunk and geometric transformer decoder (Huguet et al., 2024).
  • Molecular Conformation: Energy-based, score-based, and coupling-flow models recover equilibrium ensembles directly on Cartesian atom positions, with empirical speed-up and improved force/energy accuracies (Du et al., 2021, Midgley et al., 2023).

Evaluation metrics typically include RMSD, matching/coverage (COV/MAT), atom/molecule stability, valid/connected fractions, uniqueness, total variation of atom/bond statistics, strain energy, and task-specific geodesic error (Ryu et al., 2022, Poletukhin et al., 23 Jan 2026, Huguet et al., 2024, Wang et al., 20 Sep 2025, Ryu et al., 2023, Du et al., 2021).

Summary results:

Domain Model Data Needed Key Metric Benchmark Results
Manipulation EDF / Diffusion-EDF 5–10 demos Success rate \sim95% total, \sim1hr train
Molecule Gen. Motif Flow QM9/GEOM Atom stability (A) >>95% A, >>81% V×C, 2–10× speed
Protein Gen. FoldFlow-2 PDB/AlphaF Designability, Novelty >>97% designable, >>36% novel

6. Limitations, Data Efficiency, and Computational Complexity

SE(3)–equivariance yields dramatic reductions in sample complexity and OOD generalization. Models routinely achieve robust performance with few demonstrations or examples—5–10 for robotic manipulation; comparable contractions in molecule/protein domains (Ryu et al., 2022, Ryu et al., 2023, Poletukhin et al., 23 Jan 2026, Huguet et al., 2024).

Computational trade-offs are context-dependent:

  • Frame-based GNNs incur \sim3–5×\times speedup and memory savings over spherical harmonics-based equivariant networks (Du et al., 2021).
  • Augmented coupling flows allow %%%%29E(TX,Y)E(T|X,Y)30%%%% faster sampling and density estimation than conventional CNFs or diffusion, at the cost of EGNN passes for each layer (Midgley et al., 2023).
  • Rectified flows reduce inference steps from >>100 to 1 with superior geodesic error (Wang et al., 20 Sep 2025).

These gains are offset by the cost of group-integral sampling (e.g., MCMC/Langevin), numerical instability in deep flows, rigidity constraints in motifs, and dependence on large pretrained modalities (e.g., protein LLMs for FoldFlow-2).

7. Extensions and Future Directions

Major avenues include:

A plausible implication is that strict symmetry enforcement, combined with data-driven generative objectives (score matching, flow matching), will play an increasing role in physically-grounded AI systems for manipulation, molecular discovery, and functional structure synthesis.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to SE(3)-Equivariant Generative Modelling.