Fourier-MIONet: Neural Operator for GCS
- The paper introduces Fourier-MIONet, a neural surrogate model that embeds Fourier Neural Operator layers within a multi-input framework to map complex inputs to nonlinear PDE solutions.
- It employs dedicated branch and trunk networks with temporal inputs and U-FNO decoding to efficiently capture high-dimensional parameter-to-solution mappings for CO2 and brine flow.
- The study shows significant improvements with about 3.7M parameters, 3.5× faster training, and enhanced physical consistency over 30-year simulations compared to traditional models.
Fourier-MIONet is a Fourier-enhanced multiple-input neural operator (MIONet) designed as a neural surrogate model for multiphase flow in porous media, particularly for simulating the evolution of CO and brine in geological carbon sequestration (GCS). It achieves significant computational efficiency and generalization advantages by embedding the Fourier Neural Operator (FNO) architecture within a multiple-input operator learning framework, enabling rapid, data-efficient, and physically consistent predictions of high-dimensional parameter-to-solution maps governed by nonlinear partial differential equations (PDEs) (Jiang et al., 2023).
1. Mathematical Formulation and Operator Learning Framework
Fourier-MIONet is constructed as a parametric operator learning architecture aimed at mapping multiple spatial and scalar input fields to solution fields of interest, such as CO saturation and pressure. Given input fields over corresponding domains , the goal is to approximate the solution operator such that
where is a field over the target domain . This is achieved through the MIONet structure, where branch networks encode each input, and a trunk network encodes target coordinates : Temporal continuity is enforced by treating time as an input to a trunk net, promoting physically consistent evolution across for .
2. Network Architecture and Layer Design
Fourier-MIONet integrates FNO layers for learning global, mesh-free operator representations and U-FNO decoding to enhance expressive capacity:
- Branch Networks:
- Field branch: 2D convolutional encoder with three convolution layers (20 channels each, ReLU activations, paddings), producing latent tensors of shape (batch, , ) with retained Fourier modes.
- Scalar branch: Four-layer multilayer perceptron (MLP) with widths [10, 20, 20, 20], mapping scalars to (batch, ).
- Trunk Net: Three-layer MLP (widths 1→20→20→20) taking time and returning (batch, ).
- Fusion Operations:
- Branch merge: elementwise sum
- Branch–trunk merge: elementwise product , yielding for each time step
- Decoder (U-FNO): Four alternating FNO and U-Fourier layers in 2D, with each U-Fourier layer augmented by a small 2D U-Net. The decoder concludes with a convolution to map to the output channel.
- Parameter Count: Approximately trainable parameters—one order-of-magnitude fewer than U-FNO's .
FNO layers exploit the efficiency of the discrete Fourier transform: for a latent tensor , modes are processed with learned multipliers (acting on low-frequency modes), with the spatial representation recovered by inverse transform and pointwise linear-activation additions:
3. Governing Physical Model and Problem Setting
Fourier-MIONet addresses the nonlinear coupled PDE system describing two-phase (CO/water) flow in porous geologic formations:
- Mass conservation: For phase ,
where is porosity, is phase saturation.
- Phase flux (Darcy’s Law):
with absolute permeability , relative permeability , viscosity , and other standard parameters.
- Capillary pressure coupling: .
- Constitutive relations: and follow van Genuchten or Brooks–Corey models.
This formulation supports a wide range of input heterogeneity—including anisotropic and highly variable permeability fields, variable injection configurations, and process parameters—requiring an operator learning approach with strong generalization characteristics.
4. Training Regime, Resource Efficiency, and Performance
The network is trained on simulated datasets comprising 4,500 training instances and 500 test cases, with inputs discretized on a grid and 24 logarithmically spaced time snapshots ($1$ day to $30$ years):
- Inputs: 2D spatial fields (), scalar parameters (initial pressure , temperature , injection rate , viscosity ratio , irreducible water saturation , perforation depths).
- Loss Function: loss as in U-FNO,
balancing fieldwise error and spatial gradient error.
- Optimization: Adam optimizer (initial learning rate with decay). Batch size: (cases time snapshots) per update.
- Resource Usage Comparison:
| Model | Params | CPU Mem (GiB) | GPU Mem (GiB) | Training Time | Inference (s/case) | Speedup |
|---|---|---|---|---|---|---|
| U-FNO | $33$M | $103$ | $15.9$ | $48$ h | $0.075$ | 1 |
| Fourier-MIONet | $3.7$M | $15$ | $12.8$ | $14$ h | $0.041$ | (train), (infer) |
- Test Accuracy: , MAE , matching U-FNO.
5. Generalization, Data Efficiency, and Physical Consistency
Fourier-MIONet explicitly encodes temporal continuity in the operator by treating time as an input coordinate to the trunk network rather than as a discrete channel. This enables the model to interpolate and extrapolate smoothly over time, allowing accurate forecasting of the full 30-year CO evolution from as few as 6 (nonuniformly chosen) training snapshots (test , MAE ). This is in contrast to architectures relying on fixed time-channel discretization (e.g., U-FNO), which are unable to generalize to intermediate or unseen time points unless many snapshots are observed in training.
The approach has been validated under demanding scenarios, including permeability heterogeneity spanning four orders of magnitude, anisotropy ratios up to 10, and wide ranges of injection rates and capillary scales, consistently maintaining robust performance (Jiang et al., 2023).
6. Advantages, Limitations, and Practical Considerations
Advantages:
- One order-of-magnitude fewer parameters than U-FNO while maintaining accuracy.
- reduction in training time, faster inference.
- Substantially reduced memory footprint: 85% less CPU memory, 64% less GPU memory.
- Extreme data efficiency, supporting accurate 30-year predictions from only six data snapshots.
- Intrinsically enforces temporal continuity and physical consistency.
Limitations:
- The architecture relies on regular spatial grids to facilitate the FFT; extension to unstructured meshes is nontrivial.
- Fourier convolution’s global nature may impact scalability in extremely large or high-dimensional spatial domains.
- Hyperparameter selection (e.g., number of Fourier modes, network widths) remains application-specific.
A plausible implication is that, as with FNO-based models generally, pushing Fourier-MIONet toward very large or irregular domains may necessitate new domain decomposition or hybrid mesh strategies.
7. Broader Context and Comparison with Related Architectures
Fourier-MIONet exemplifies the integration of operator learning (MIONet) and Fourier-domain modeling (FNO), aligning with recent advances in mesh-free, global-receptive-field deep architectures for scientific machine learning. In operator learning, FNO layers enable efficient learning of parametric solution maps for PDE-governed systems, and the multi-input/trunk design allows explicit inclusion of both field and scalar process variables.
The architectural principles of Fourier-MIONet are applicable to a broad class of problems requiring solution of high-dimensional, parameterized PDEs—extending beyond geological carbon sequestration to possible applications in subsurface hydrology, reactive transport, and time-continuous simulation surrogates (Jiang et al., 2023).
Fourier neural operators have also begun to demonstrate potential in physical-layer modeling for complex electromagnetic and communications systems, highlighting the generality of the core convolutional operator-learning paradigm (Xiao et al., 6 Oct 2025). However, Fourier-MIONet's contributions are characterized by its tailored multi-input framework, physical consistency over time, and demonstrated empirical gains in data efficiency and computational resource usage for carbon storage simulation.