Papers
Topics
Authors
Recent
Search
2000 character limit reached

LD-GCN: Latent Dynamics for Model Reduction

Updated 19 January 2026
  • LD-GCN is a data-driven architecture that uses a latent neural ODE and graph convolutional decoder to achieve model order reduction on unstructured meshes.
  • The method employs a two-branch design where the latent dynamics evolve causally with explicit Euler integration and are decoded to high-dimensional solution fields.
  • Empirical results demonstrate LD-GCN’s ability to reduce errors compared to prior approaches, with theoretical guarantees and effective zero-shot prediction via latent interpolation.

The Latent Dynamics Graph Convolutional Network (LD-GCN) is a data-driven, encoder-free neural architecture tailored for nonlinear model order reduction (MOR) of parameterized and time-dependent partial differential equations (PDEs) on unstructured meshes. LD-GCN combines a global, low-dimensional latent representation of system dynamics with graph convolutional decoding to reconstruct high-dimensional solution fields, advancing the latent state causally in time through a neural ODE. This approach explicitly conditions the reduced dynamics on physical and geometric parameters, achieves interpretability of latent trajectories, provides zero-shot prediction via latent interpolation, and supports rigorous theoretical guarantees under precise mathematical assumptions (Tomada et al., 16 Jan 2026).

1. Architectural Fundamentals

LD-GCN employs an encoder-free construction in which the latent state s(t;μ)Rn\bm s(t;\bm\mu)\in\mathbb R^n is initialized as s(0;μ)=0\bm s(0;\bm\mu)=\bm0, without mapping the field into latent space via a separate encoder. The network advances the latent state in time through a neural ordinary differential equation (NODE), leveraging a two-branch design:

  • Latent-dynamics branch ("NN-dyn"): Implements a fully connected neural ODE, computing s˙(t)=NNdyn(t,s(t),μ(t))\dot{\bm s}(t)=\mathrm{NN}_{\rm dyn}(t,\bm s(t),\bm\mu(t)) with explicit Euler integration.
  • Graph-convolutional decoder ("NN-dec"): Given (t,s(t),μ(t))(t,\bm s(t),\bm\mu(t)), reconstructs the full-order field usim(t;μ)RNh×du\bm u_\text{sim}(t;\bm\mu)\in\mathbb R^{N_h\times d_u} on mesh M(V,E)\mathcal M(\mathcal V,\mathcal E) using a small MLP bottleneck, two MoNet-based graph convolutional layers with nhcn_{\rm hc} hidden channels and ELU activations, and a final linear read-out per node.

This encoder-free causal formulation ensures that the evolution of the latent state is entirely governed by the NODE and external parameters, allowing for direct interpretability and time extrapolation.

2. Latent Dynamics Formulation

LD-GCN models system dynamics in the latent space via a continuous neural ODE,

ddts(t)=NNdyn(t,  s(t),  μ(t)),s(0)=0,\frac{d}{dt}\,\bm s(t)=\mathrm{NN}_{\rm dyn}\bigl(t,\;\bm s(t),\;\bm\mu(t)\bigr),\quad\bm s(0)=\bm0,

with high-dimensional solution field decoded as

usim(t;μ)=NNdec(t,  s(t),  μ(t)).\bm u_\mathrm{sim}(t;\bm\mu)=\mathrm{NN}_{\rm dec}\bigl(t,\;\bm s(t),\;\bm\mu(t)\bigr).

In practical implementations, time is discretized using explicit Euler steps:

s(tj+1)=s(tj)+ΔtNNdyn(tj,s(tj),μ(tj)).\bm s(t_{j+1})=\bm s(t_j)+\Delta t\,\mathrm{NN}_{\rm dyn}(t_j,\bm s(t_j),\bm\mu(t_j)).

Both temporal (tt) and parameter (μ(t)\bm\mu(t)) dependencies are explicitly input to NNdyn\mathrm{NN}_{\rm dyn} (and optionally NNdec\mathrm{NN}_{\rm dec}), enabling the latent state to respond to external forcing and parameter variations.

3. Mathematical Guarantees

The LD-GCN construction admits a universal approximation result for encoder-free architectures (Corollary 3.1), under the assumption of perfect latent embedding:

  • There exists a sufficiently expressive latent NODE NNdyn\mathcal{NN}_{\rm dyn} and graph decoder NNdec\mathcal{NN}_{\rm dec} such that, for any ε>0\varepsilon>0, integration step Δt\Delta t^*, and all ΔtΔt\Delta t\leq\Delta t^*, the network can approximate the full-order solution uh(;μ)\bm u_h(\cdot;\bm\mu) within ε\varepsilon at all time steps and parameter points:

supk=1,,Nt  uh(tk;μ)usim(tk;μ)ε,\sup_{k=1,\dots,N_t}\;\|\bm u_h(t_k;\bm\mu)-\bm u_\text{sim}(t_k;\bm\mu)\|\leq\varepsilon,

where the simulated field is computed as

usim(tk;μ)=NNdec(Δtj=0kωjNNdyn(tj,μ)).\bm u_\mathrm{sim}(t_k;\bm\mu)=\mathcal{NN}_\mathrm{dec}\Bigl(\Delta t\sum_{j=0}^k\omega_j\,\mathcal{NN}_\mathrm{dyn}(t_j,\bm\mu)\Bigr).

Assumptions: fixed initial conditions, smooth decoder existence, Lipschitz continuity, and time-constant signals (extensions to fully time-varying forcing remain open).

4. Training and Optimization Protocol

LD-GCN is trained using a composite loss: L(w)=Lerr(uh,usim)+λw1,\mathcal L(\bm w)=\mathcal L_{\rm err}(\bm u_h,\bm u_\text{sim})+\lambda\|\bm w\|_1, where Lerr\mathcal L_{\rm err} denotes mean squared error (MSE), supplemented with a directional-alignment term Lε\mathcal L_\varepsilon (weighted by δ=101\delta=10^{-1}) for Navier–Stokes cases. L1L^1 weight decay is applied with strength λ=105\lambda=10^{-5}.

Training proceeds via:

  • Initial Adam optimization (200–1500 epochs)
  • Final refinement using L-BFGS (200 steps)
  • Backpropagation through the latent-ODE integration loop

Data requirements entail assembling a training set TtrainI×P\mathcal T_{\rm train}\subset I\times P; the latter portion (typically last 25–50%) of each trajectory is withheld to test time extrapolation. Solution snapshots are affinely normalized per node and no scaling is applied to tt or μ\bm\mu.

5. Zero-Shot Prediction and Latent Interpolation

LD-GCN supports zero-shot prediction by constructing explicit interpolants of latent trajectories s~(t;μ)s(t;μ)\tilde{\bm s}(t;\bm\mu)\approx\bm s(t;\bm\mu) over the training mesh, which are decoded via the trained NN-dec. Two interpolation methods are employed:

  • Degree-1 multivariate splines (support interpolation only)
  • Gaussian process regression (GPR) with Matérn ν=1.5\nu=1.5 kernel (supports extrapolation in tt and μ\bm\mu)

Proposition 3.3 establishes the error bound for reconstructed fields using latent interpolation: uh(t;μ)uinterp(t;μ)Lδ(m)+ε(t;μ),\|\bm u_h(t;\bm\mu)-\bm u_\text{interp}(t;\bm\mu)\|\leq L\,\delta(m)+\varepsilon(t;\bm\mu), where LL is the Lipschitz constant of NN-dec and δ(m)\delta(m) the maximal discrepancy in latent space interpolation. As δ(m)0\delta(m)\to0, the surrogate approaches full LD-GCN accuracy.

Two practical strategies for handling unseen (t,μ)(t,\bm\mu) points are outlined:

  1. Integrate-then-interpolate: advance all latent trajectories then interpolate.
  2. Interpolate-then-extrapolate: build a GPR directly on latent space and query new input pairs.

6. Empirical Results on Benchmark Problems

LD-GCN has been evaluated across several computational mechanics benchmarks:

Benchmark Mesh Nodes (NhN_h) Latent Dim (nn) Mean Rel. Error Max Rel. Error Prior Approach (GCA) Error
SA (advection) 1472 3 7.86×1037.86 \times 10^{-3} 5.26×1025.26 \times 10^{-2} 1.87×1021.87 \times 10^{-2} (mean)
MH (moving hole) 1352 15 8.81×1038.81 \times 10^{-3} 1.02×1011.02 \times 10^{-1} 3.54×1023.54 \times 10^{-2} (mean)
Lid cavity (NS) 10024 3 NRMSE 6.79×1036.79 \times 10^{-3} LD-Net NRMSE 1.39×1031.39 \times 10^{-3} (n=10)(n=10)
Coandă effect (NS) 4.57×1034.57 \times 10^{-3} 3.62×1023.62 \times 10^{-2}

For advection–diffusion benchmarks (SA and MH), LD-GCN achieves lower relative errors over GCA, demonstrates robust handling of geometric parameters, and encodes smooth, interpretable latent trajectories. In Navier–Stokes problems (lid cavity, Coandă effect), LD-GCN reproduces complex phenomena including bifurcation diagrams and tracks time-dependent boundary forcing via latent variables.

Zero-shot predictions via GPR and splines on the SA benchmark yield comparable error profiles; GPR affords natural extrapolation capabilities.

7. Implementation Specifications

Reference code is available at https://github.com/lorenzotomada/ld-gcn-rom.

Architectural details for NN-dyn (latent dynamics MLP):

  • SA/MH/Coandă: five layers with widths [n+2,80,,80,n][n+2, 80,\dots, 80, n] or [n+1,50,50,n][n+1, 50,50, n]
  • Lid cavity: seven layers [n+2,100,100,100,100,100,n][n+2, 100, 100, 100, 100, 100, n]
  • Activation: tanh

Graph decoder (NN-dec):

  • Two-layer MLP projecting (t,s,μ)(t,\bm s,\bm\mu) to RNh×nhc\mathbb R^{N_h\times n_{\rm hc}}; nhc=2n_{\rm hc}=2
  • Two MoNet graph-convolutional layers with ELU activations and skip-connections
  • Final linear layer for dud_u output per node

Hyperparameters:

  • Learning rate 10310^{-3}, L1L^1 weight decay λ=105\lambda=10^{-5}
  • Navier–Stokes extra loss weight δ=101\delta=10^{-1}, ε=104\varepsilon=10^{-4}
  • Optimizer: Adam (200–1500 epochs) to L-BFGS (200 steps)
  • Explicit Euler Δt\Delta t chosen per problem, e.g., 0.5 (SA/MH/Coandă), 0.1 (lid cavity)
  • Pre/post-processing: per-node normalization; no scaling of tt or μ\bm\mu

This architecture delivers interpretable global latent dynamics, non-intrusive handling of parameterized PDEs on unstructured meshes, direct causal time evolution and extrapolation, and scalability relative to full-order models and prior graph autoencoder-based approaches (Tomada et al., 16 Jan 2026).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Latent Dynamics Graph Convolutional Network (LD-GCN).