Papers
Topics
Authors
Recent
Search
2000 character limit reached

DA-SHRED: Shallow Recurrent Decoder Assimilation

Updated 3 December 2025
  • The paper presents a latent assimilation framework that compresses high-dimensional states into a low-dimensional space for real-time reconstruction.
  • It combines a shallow encoder-decoder with a recurrent model and Kalman-style updates to integrate sparse sensor data and simulation proxies.
  • Sparse regression (SINDy) identifies missing dynamical terms, achieving a significant reduction in RMSE and bridging the SIM2REAL gap.

Data Assimilation with a SHallow REcurrent Decoder (DA-SHRED) is a machine learning framework designed to integrate sparse sensor data with computational simulation models for high-dimensional, spatiotemporal physical systems. It operates by embedding the full system state into a low-dimensional latent space, enabling real-time reconstruction and discrepancy modeling between model predictions and experimental measurements. The methodology addresses the simulation-to-real (SIM2REAL) gap introduced by unmodeled physics and parameter misspecification, providing both assimilation and identification of missing dynamics through sparse-regression in the latent space (Bao et al., 1 Dec 2025).

1. Problem Formulation and Mathematical Framework

DA-SHRED considers a high-dimensional system state xt∈Rnx_t \in \mathbb{R}^n evolving under unknown real physics. Available resources are sparse point-sensor measurements yt∈Rpy_t \in \mathbb{R}^p and a reduced simulation proxy NN that approximates the true system dynamics,  x˙=N(x,t) \,\dot{x} = N(x, t)\,. Observations are modeled as  yt=Hxt+ηt \,y_t = H x_t + \eta_t\,, with H∈Rp×nH \in \mathbb{R}^{p \times n} a known linear observation operator and ηt\eta_t measurement noise.

The dual objectives are:

  • Assimilate incoming measurements yty_t into a reduced latent representation zt∈Rr  (r≪n)z_t \in \mathbb{R}^r\,\,(r \ll n) to reconstruct the full state x^t≈xt\hat{x}_t \approx x_t in real time.
  • Discover missing or unmodeled dynamics yt∈Rpy_t \in \mathbb{R}^p0 such that the true dynamics are yt∈Rpy_t \in \mathbb{R}^p1.

The framework employs:

  • A shallow encoder yt∈Rpy_t \in \mathbb{R}^p2, yt∈Rpy_t \in \mathbb{R}^p3
  • A recurrent latent model yt∈Rpy_t \in \mathbb{R}^p4, yt∈Rpy_t \in \mathbb{R}^p5
  • A shallow decoder yt∈Rpy_t \in \mathbb{R}^p6, yt∈Rpy_t \in \mathbb{R}^p7

Superscripts yt∈Rpy_t \in \mathbb{R}^p8 denote forecast and analysis, respectively.

2. SHRED Architecture and Implementation

SHRED employs an encoder-decoder sequence without a traditional autoencoder inverse. The encoder yt∈Rpy_t \in \mathbb{R}^p9 is either a single linear layer or a small MLP mapping full-state snapshots into a low-dimensional latent space. The decoder NN0 is shallow, typically a single linear layer (possibly with a nonlinearity), that reconstructs the full grid from latent codes.

Temporal dynamics in latent space are captured via NN1, usually instantiated as an LSTM or small RNN:

NN2

For simulation-only training, reconstruction is enforced via:

  • NN3
  • NN4
  • NN5

with mean-square error minimization over simulated trajectory NN6.

3. Latent Data Assimilation Procedure

At each time step, the procedure executes:

  • Forecast: NN7
  • Innovation: NN8
  • Analysis update: NN9, with  xË™=N(x,t) \,\dot{x} = N(x, t)\,0 as the gain matrix mapping innovations to latent corrections.

Post-update, full-state is decoded:  x˙=N(x,t) \,\dot{x} = N(x, t)\,1, supporting comparisons in sensor or full-domain space.

4. Discrepancy Modeling via Sparse Identification

DA-SHRED includes a sparse regression stage to model missing physics in latent space using SINDy (Sparse Identification of Nonlinear Dynamics). For an assimilated latent trajectory  x˙=N(x,t) \,\dot{x} = N(x, t)\,2, finite-difference approximations yield  x˙=N(x,t) \,\dot{x} = N(x, t)\,3.

Missing latent dynamics are hypothesized to be sparse in a dictionary  x˙=N(x,t) \,\dot{x} = N(x, t)\,4 of candidate nonlinear functions. SINDy regression solves:

 x˙=N(x,t) \,\dot{x} = N(x, t)\,5

where  x˙=N(x,t) \,\dot{x} = N(x, t)\,6,  x˙=N(x,t) \,\dot{x} = N(x, t)\,7, and nonzero entries of  x˙=N(x,t) \,\dot{x} = N(x, t)\,8 identify active nonlinearities. Physical corrections  x˙=N(x,t) \,\dot{x} = N(x, t)\,9 are projected back to physical space via the decoder basis.

5. Training Objectives and Joint Optimization

The overall learning problem jointly tunes:

  • Encoder-decoder parameters  yt=Hxt+ηt \,y_t = H x_t + \eta_t\,0
  • Latent recurrent model  yt=Hxt+ηt \,y_t = H x_t + \eta_t\,1
  • Assimilation gains  yt=Hxt+ηt \,y_t = H x_t + \eta_t\,2
  • SINDy coefficients  yt=Hxt+ηt \,y_t = H x_t + \eta_t\,3

The main loss components are:

  1. Simulation-only reconstruction:

 yt=Hxt+ηt \,y_t = H x_t + \eta_t\,4

  1. Data-assimilation loss:

 yt=Hxt+ηt \,y_t = H x_t + \eta_t\,5

  1. Discrepancy (SINDy) loss:

 yt=Hxt+ηt \,y_t = H x_t + \eta_t\,6

Combined optimization:

 yt=Hxt+ηt \,y_t = H x_t + \eta_t\,7

with  yt=Hxt+ηt \,y_t = H x_t + \eta_t\,8 as weighting hyperparameters.

6. Representative Test Cases and Quantitative Evaluation

Empirical evaluations cover:

  • 2D damped Kuramoto–Sivashinsky (KS) system on  yt=Hxt+ηt \,y_t = H x_t + \eta_t\,9
  • 2D Kolmogorov flow (Navier–Stokes with sinusoidal forcing)
  • 2D Gray–Scott reaction–diffusion system
  • 1D rotating detonation engine (RDE) model

Metrics include full-field RMSE, H∈Rp×nH \in \mathbb{R}^{p \times n}0, and sensor RMSE, H∈Rp×nH \in \mathbb{R}^{p \times n}1.

Key outcomes:

  • DA-SHRED achieves %%%%52yt∈Rpy_t \in \mathbb{R}^p053%%%% reduction in full-field RMSE within H∈Rp×nH \in \mathbb{R}^{p \times n}4–H∈Rp×nH \in \mathbb{R}^{p \times n}5 time units, compared to the simulation-only proxy.
  • Robust correction with few sensors: H∈Rp×nH \in \mathbb{R}^{p \times n}6 simulated, H∈Rp×nH \in \mathbb{R}^{p \times n}7–H∈Rp×nH \in \mathbb{R}^{p \times n}8 real.
  • SINDy module precisely recovers missing dynamical terms, e.g., H∈Rp×nH \in \mathbb{R}^{p \times n}9 in KS, ηt\eta_t0 in Kolmogorov flow, ηt\eta_t1 in Gray–Scott, ηt\eta_t2 in RDE.

7. Synthesis, Practical Implications, and Extensions

DA-SHRED unites three major components:

  1. Efficient compression of high-dimensional PDE states via a shallow encoder–recurrent–decoder structure yielding a compact latent representation amenable to rapid computation.
  2. Latent assimilation loop implementing Kalman-style updates for incorporating sparse, noisy sensor data in real time.
  3. Physics-informed discrepancy inference through sparse regression (SINDy) in latent coordinates, facilitating explicit identification of missing or uncaptured processes.

This synergy supports robust closure of the SIM2REAL gap—empirically %%%%63yt∈Rpy_t \in \mathbb{R}^p064%%%% RMSE reduction compared with pure simulation—and enables interpretable extraction of dynamical corrections (Bao et al., 1 Dec 2025). The approach generalizes to a variety of physical systems and sensor modalities, providing a scalable, computationally efficient framework for digital-twin deployment, model correction, and high-fidelity state reconstruction.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Data Assimilation with a SHallow REcurrent Decoder (DA-SHRED).