Papers
Topics
Authors
Recent
Search
2000 character limit reached

WeatherFormer: Empowering Global Numerical Weather Forecasting with Space-Time Transformer

Published 21 Sep 2024 in cs.AI, cs.LG, and physics.ao-ph | (2409.16321v1)

Abstract: Numerical Weather Prediction (NWP) system is an infrastructure that exerts considerable impacts on modern society.Traditional NWP system, however, resolves it by solving complex partial differential equations with a huge computing cluster, resulting in tons of carbon emission. Exploring efficient and eco-friendly solutions for NWP attracts interest from AI and earth science communities. To narrow the performance gap between the AI-based methods and physic predictor, this work proposes a new transformer-based NWP framework, termed as WeatherFormer, to model the complex spatio-temporal atmosphere dynamics and empowering the capability of data-driven NWP. WeatherFormer innovatively introduces the space-time factorized transformer blocks to decrease the parameters and memory consumption, in which Position-aware Adaptive Fourier Neural Operator (PAFNO) is proposed for location sensible token mixing. Besides, two data augmentation strategies are utilized to boost the performance and decrease training consumption. Extensive experiments on WeatherBench dataset show WeatherFormer achieves superior performance over existing deep learning methods and further approaches the most advanced physical model.

Citations (1)

Summary

  • The paper demonstrates that WeatherFormer narrows the gap between AI-based methods and physical predictors using a novel space-time transformer architecture.
  • The methodology integrates space-time factorized blocks and Fourier transformations to reduce computational complexity while enhancing prediction accuracy.
  • Experiments on WeatherBench show superior RMSE and ACC metrics, suggesting significant energy savings and improved forecast reliability.

WeatherFormer: Empowering Global Numerical Weather Forecasting with Space-Time Transformer

Introduction

The paper introduces WeatherFormer, a transformer-based framework designed to enhance global numerical weather prediction (NWP) using a space-time factorized architecture. Traditional NWP systems rely on solving complex partial differential equations, which are computationally expensive and lead to significant carbon emissions. WeatherFormer seeks to narrow the performance gap between AI-based methods and physical predictors by modeling the dynamics of the atmosphere through a combination of space-time factorized transformer blocks and the Position-aware Adaptive Fourier Neural Operator (PAFNO).

Methodology

Framework Overview

WeatherFormer leverages the spatio-temporal relationships in weather data to improve prediction accuracy. The framework uses a sequence of weather states as input and outputs the future states over multiple time steps. It divides input data into tokens that are processed by a series of SF-Blocks containing both spatial and temporal mixers (Figure 1). These mixers employ a Fourier-based approach to handle high-dimensional data efficiently. Figure 1

Figure 1: Overview of WeatherFormer, illustrating its architecture leveraging space-time factorized transformer blocks and PAFNO for efficient weather state prediction.

Factorized Space-Time Block

The SF-Block addresses the challenge of processing spatio-temporal data by separating spatial and temporal relationships, reducing the computational complexity. This block uses Fourier transformations to process token information, enabling the model to predict future weather states by considering historical data sequences.

Position Aware Adaptive Fourier Neural Operator (PAFNO)

PAFNO enhances the standard Fourier Neural Operator by introducing position-aware coefficients, allowing the model to utilize spatial information effectively. This approach ensures that the model accounts for spatial dependencies critical to accurate weather predictions, offering improvements over existing data-driven NWP methods.

Experiments and Results

The performance of WeatherFormer was evaluated on the WeatherBench dataset, showing superior results compared to both traditional physics-based models and contemporary deep learning approaches. WeatherFormer demonstrated a closer performance to advanced physical models while maintaining a lower computational footprint.

Performance metrics like RMSE and ACC were used to benchmark against existing methods, highlighting WeatherFormer's efficient and effective prediction capabilities for key meteorological variables like Z500, T850, and T2M. The model also successfully mitigated long-term error accumulation with noise augmentation strategies.

Conclusion

WeatherFormer presents a robust approach to data-driven weather prediction, showcasing potential for significant energy savings compared to traditional NWP methods while achieving competitive accuracy levels. The integration of space-time transformers and adaptive Fourier neural operators positions WeatherFormer as a promising alternative for future weather prediction systems, particularly as the field continues to advance towards more sustainable and efficient methodologies. Future work may focus on further refining these techniques and exploring additional applications in Earth system science and beyond.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Authors (4)

Collections

Sign up for free to add this paper to one or more collections.