LD-GCN: Latent Dynamics for Model Reduction
- LD-GCN is a data-driven architecture that uses a latent neural ODE and graph convolutional decoder to achieve model order reduction on unstructured meshes.
- The method employs a two-branch design where the latent dynamics evolve causally with explicit Euler integration and are decoded to high-dimensional solution fields.
- Empirical results demonstrate LD-GCN’s ability to reduce errors compared to prior approaches, with theoretical guarantees and effective zero-shot prediction via latent interpolation.
The Latent Dynamics Graph Convolutional Network (LD-GCN) is a data-driven, encoder-free neural architecture tailored for nonlinear model order reduction (MOR) of parameterized and time-dependent partial differential equations (PDEs) on unstructured meshes. LD-GCN combines a global, low-dimensional latent representation of system dynamics with graph convolutional decoding to reconstruct high-dimensional solution fields, advancing the latent state causally in time through a neural ODE. This approach explicitly conditions the reduced dynamics on physical and geometric parameters, achieves interpretability of latent trajectories, provides zero-shot prediction via latent interpolation, and supports rigorous theoretical guarantees under precise mathematical assumptions (Tomada et al., 16 Jan 2026).
1. Architectural Fundamentals
LD-GCN employs an encoder-free construction in which the latent state is initialized as , without mapping the field into latent space via a separate encoder. The network advances the latent state in time through a neural ordinary differential equation (NODE), leveraging a two-branch design:
- Latent-dynamics branch ("NN-dyn"): Implements a fully connected neural ODE, computing with explicit Euler integration.
- Graph-convolutional decoder ("NN-dec"): Given , reconstructs the full-order field on mesh using a small MLP bottleneck, two MoNet-based graph convolutional layers with hidden channels and ELU activations, and a final linear read-out per node.
This encoder-free causal formulation ensures that the evolution of the latent state is entirely governed by the NODE and external parameters, allowing for direct interpretability and time extrapolation.
2. Latent Dynamics Formulation
LD-GCN models system dynamics in the latent space via a continuous neural ODE,
with high-dimensional solution field decoded as
In practical implementations, time is discretized using explicit Euler steps:
Both temporal () and parameter () dependencies are explicitly input to (and optionally ), enabling the latent state to respond to external forcing and parameter variations.
3. Mathematical Guarantees
The LD-GCN construction admits a universal approximation result for encoder-free architectures (Corollary 3.1), under the assumption of perfect latent embedding:
- There exists a sufficiently expressive latent NODE and graph decoder such that, for any , integration step , and all , the network can approximate the full-order solution within at all time steps and parameter points:
where the simulated field is computed as
Assumptions: fixed initial conditions, smooth decoder existence, Lipschitz continuity, and time-constant signals (extensions to fully time-varying forcing remain open).
4. Training and Optimization Protocol
LD-GCN is trained using a composite loss: where denotes mean squared error (MSE), supplemented with a directional-alignment term (weighted by ) for Navier–Stokes cases. weight decay is applied with strength .
Training proceeds via:
- Initial Adam optimization (200–1500 epochs)
- Final refinement using L-BFGS (200 steps)
- Backpropagation through the latent-ODE integration loop
Data requirements entail assembling a training set ; the latter portion (typically last 25–50%) of each trajectory is withheld to test time extrapolation. Solution snapshots are affinely normalized per node and no scaling is applied to or .
5. Zero-Shot Prediction and Latent Interpolation
LD-GCN supports zero-shot prediction by constructing explicit interpolants of latent trajectories over the training mesh, which are decoded via the trained NN-dec. Two interpolation methods are employed:
- Degree-1 multivariate splines (support interpolation only)
- Gaussian process regression (GPR) with Matérn kernel (supports extrapolation in and )
Proposition 3.3 establishes the error bound for reconstructed fields using latent interpolation: where is the Lipschitz constant of NN-dec and the maximal discrepancy in latent space interpolation. As , the surrogate approaches full LD-GCN accuracy.
Two practical strategies for handling unseen points are outlined:
- Integrate-then-interpolate: advance all latent trajectories then interpolate.
- Interpolate-then-extrapolate: build a GPR directly on latent space and query new input pairs.
6. Empirical Results on Benchmark Problems
LD-GCN has been evaluated across several computational mechanics benchmarks:
| Benchmark | Mesh Nodes () | Latent Dim () | Mean Rel. Error | Max Rel. Error | Prior Approach (GCA) Error |
|---|---|---|---|---|---|
| SA (advection) | 1472 | 3 | (mean) | ||
| MH (moving hole) | 1352 | 15 | (mean) | ||
| Lid cavity (NS) | 10024 | 3 | NRMSE | — | LD-Net NRMSE |
| Coandă effect (NS) | — | — | — |
For advection–diffusion benchmarks (SA and MH), LD-GCN achieves lower relative errors over GCA, demonstrates robust handling of geometric parameters, and encodes smooth, interpretable latent trajectories. In Navier–Stokes problems (lid cavity, Coandă effect), LD-GCN reproduces complex phenomena including bifurcation diagrams and tracks time-dependent boundary forcing via latent variables.
Zero-shot predictions via GPR and splines on the SA benchmark yield comparable error profiles; GPR affords natural extrapolation capabilities.
7. Implementation Specifications
Reference code is available at https://github.com/lorenzotomada/ld-gcn-rom.
Architectural details for NN-dyn (latent dynamics MLP):
- SA/MH/Coandă: five layers with widths or
- Lid cavity: seven layers
- Activation: tanh
Graph decoder (NN-dec):
- Two-layer MLP projecting to ;
- Two MoNet graph-convolutional layers with ELU activations and skip-connections
- Final linear layer for output per node
Hyperparameters:
- Learning rate , weight decay
- Navier–Stokes extra loss weight ,
- Optimizer: Adam (200–1500 epochs) to L-BFGS (200 steps)
- Explicit Euler chosen per problem, e.g., 0.5 (SA/MH/Coandă), 0.1 (lid cavity)
- Pre/post-processing: per-node normalization; no scaling of or
This architecture delivers interpretable global latent dynamics, non-intrusive handling of parameterized PDEs on unstructured meshes, direct causal time evolution and extrapolation, and scalability relative to full-order models and prior graph autoencoder-based approaches (Tomada et al., 16 Jan 2026).