FourCastNet3: Geometric Weather Forecasting
- FourCastNet3 is a purely convolutional geometric ML model that leverages spherical operator blocks to forecast global weather probabilistically.
- The model employs advanced ensemble techniques with hidden Markov processes and spectral fidelity metrics to achieve near-unity spread–skill ratios and robust calibration.
- FCN3 delivers exceptional computational performance by generating 90-day, 0.25° global forecasts in under 20 seconds on a single GPU, enabling real-time meteorological applications.
FourCastNet3 (FCN3) is a purely convolutional, geometric machine learning model for probabilistic, global, medium- to long-range weather forecasting. It approximates the evolution operator on the sphere , integrating spherical geometry into both model architecture and phyiscally consistent ensemble prediction. FCN3 delivers forecasting accuracy that surpasses leading conventional ensemble models and rivals diffusion-based ML methods, while achieving exceptional computational efficiency—enabling 90-day, 6-hourly global forecasts at 0.25° resolution in under 20 seconds on a single GPU. The system is robust to initialization sources and maintains spectral and ensemble skill at extended forecasts up to 60 days. These advances are realized through a cascade of spherical neural-operator blocks, scalable domain-decomposition-based training, and novel statistical ensemble generation, establishing FCN3's role as a candidate for practical meteorological and climate applications, including risk assessment and early warning systems (Bonev et al., 16 Jul 2025, Gupta et al., 25 Jan 2026).
1. Geometric Model Architecture and Spherical Operator Design
FCN3's architecture consists of three stages—an encoder, a stack of spherical neural-operator (“processor”) blocks, and a decoder. Primary inputs include atmospheric levels, surface variables, auxiliary fields, and stochastic latent noise, presented on a 0.25° equiangular latitude–longitude grid ( points); internal computations are conducted on a coarser Gaussian grid.
The encoder and decoder employ grouped discrete-continuous (DISCO) convolutions combined with bilinear interpolation for dimension changes, delaying channel mixing to deep layers to avoid premature blending of physically distinct variables.
The processor comprises 10 ConvNeXt-style residual blocks: 2 global (spectral) blocks and 8 local blocks. Each block contains a spherical convolution—either global (spectral) or local (DISCO)—followed by a two-layer MLP with GeLU activation and learnable LayerScale residual weights. Global convolutions invoke the convolution theorem on :
with spherical-harmonic coefficient update:
Local DISCO convolutions utilize compactly supported Morlet-like wavelet basis functions to yield sparse neighborhood connectivity. Spherical harmonic transforms (SHTs) and inverse SHTs efficiently map between grid samples and truncated spherical coefficients with scaling.
Spherical power spectra are computed as:
2. Probabilistic Ensemble Forecasting and Calibration
FCN3 employs a probabilistic hidden-Markov ensemble model, predicting next states as:
where represents latent noise drawn from a mixture of spherical diffusion processes:
with , , and .
Ensembles are generated by sampling for each member , yielding that model . Calibration is assessed via ensemble mean, variance, spread-skill ratio, rank histograms, and Continuous Ranked Probability Score (CRPS):
FCN3 achieves near-unity spread–skill ratios and flat rank histograms over extended periods, with CRPS matching or outperforming state-of-the-art ML and NWP ensemble systems.
3. Scalable Distributed Training Paradigm
Training involves deterministic pretraining (MSE loss on ), then ensemble training minimizing a composite loss combining pointwise CRPS and multiscale spectral CRPS (to enforce correct spatial power spectra). FCN3 introduces a hybrid parallelism strategy:
- Model parallelism: domain decomposition along latitude; each GPU processes a latitude slab; distributed SHTs and DISCO ops.
- Data parallelism: across ensemble members and batch samples; gradient sync via MPI all-reduce (data-parallel) and slabwise communication (model-parallel).
- Implementation: all operations are defined in the Makani framework (torch-harmonics), enabling synchronous, linearly scaling training on up to 2048 H100 GPUs, with training times of 78 h (first stage, 1024 H100), 15 h (second stage, 512 A100), and 8 h (finetuning, 256 H100).
4. Forecast Skill, Spectral Fidelity, and Robustness
FCN3 achieves high quantitative skill in year-long 2020 tests at 12-h lead-time resolution. Its 50-member ensemble:
- CRPS & RMSE: Outperforms ECMWF IFS-ENS on of channels and lead times; matches GenCast’s CRPS at double the temporal resolution.
- Calibration: Consistently flat rank histograms; spread–skill ratios near unity.
- Spectral performance: Preserves spherical-harmonic power law slopes (–3 at synoptic, –5/3 at mesoscale) with relative error up to .
- Rollout stability: Extended forecasts up to 60 days exhibit no large-scale blowups; spectral slopes remain invariant.
Case studies (e.g., Storm Dennis, February 2020) confirm accurate storm track, amplitude, and spectra for up to 30 days.
FCN3's robustness to initialization from ML-based data assimilation schemes, such as HealDA, is quantitatively validated (Gupta et al., 25 Jan 2026). When initialized with HealDA rather than ERA5 analyses:
- FCN3 loses only –18 h effective lead time in forecast skill.
- Forecast error growth exponent remains unchanged (–).
- The penalty stems mainly from larger large-scale () initial errors in HealDA; small-scale variance is rapidly regenerated.
- FCN3 is insensitive to modest IC grid smoothing, observation timing, and sensor dropout (few hour shifts in performance).
Table 1. FCN3 RMSE (Z500 in m²/s², T850 in K, U500 in m/s)
| Init | IC RMSE | RMSE @ 1 day | RMSE @ 5 days |
|---|---|---|---|
| ERA5 | 25; 0.34; 0.74 | 47; 0.77; 1.46 | 280; 1.62; 4.22 |
| HealDA | 47; 0.70; 1.68 | 67; 0.78; 1.98 | 325; 1.76; 4.71 |
The h lead penalty with HealDA initialization suggests rapid error saturation is dominated by large-scale analysis errors, with no impact on the dynamical instability exponent.
5. Computational Efficiency and Large-Scale Applicability
FCN3 is computationally efficient, producing a 90-day, 6-hourly global forecast at 0.25° resolution for 50 ensemble members in s on a single NVIDIA H100 GPU—about faster than GenCast and faster than IFS-ENS. Plug-and-play initialization with ML DA systems such as HealDA can be achieved in s per analysis (Gupta et al., 25 Jan 2026).
The low inference cost, scalability to large ensembles (1000+ members), and generation of well-calibrated probabilistic outputs enable applications in rare-event risk assessment, subseasonal and in-situ forecasting, and real-time ensemble updates. Open-source frameworks (Makani, torch-harmonics) support reproducibility and community involvement.
A plausible implication is that further improvements in ML DA for weather forecasting will require reduction of large-scale analysis error—potentially achievable through enriched priors, expanded datasets, or novel network regularization targeting synoptic structures.
6. Connections to Data Assimilation and ML Forecast Ecosystem
State-of-the-art ML DA systems such as HealDA produce initial atmospheric states using only satellite and conventional observations, without reliance on NWP forecast backgrounds. FCN3 can be initialized by these analyses with only modest lead-time skill loss—a horizontal shift of –18 h relative to NWP-based initial states. The error-growth rate and robustness to smoothing indicate FCN3’s insensitivity to IC details, provided that large-scale structural errors are minimized.
Routine forecast verification demonstrates that scoring protocols (cycle timings, grid resolutions, sensor outages) can shift apparent skill by 3–24 h, underscoring the importance of consistent evaluation methodologies for ML weather models. FCN3’s compatibility with ML DA systems untethers ensemble prediction from expensive legacy infrastructure, reducing operational barriers for global ML-driven weather prediction (Gupta et al., 25 Jan 2026).
7. Future Directions and Implications
Advances in FCN3 and its ecosystem point toward future research on further geometric model refinement, improved spectral regularization, and advanced methods for mitigating large-scale initialization errors. The capacity for large ensembles, rapid inference, and accurate spectral representation make FCN3 a strong candidate for operational meteorology and climate systems. Continued progress in ML DA and further open-source development are expected to strengthen the reliability and utility of purely data-driven global weather models (Bonev et al., 16 Jul 2025, Gupta et al., 25 Jan 2026).