Papers
Topics
Authors
Recent
Search
2000 character limit reached

Predicting waves in fluids with deep neural network

Published 17 Jan 2022 in physics.flu-dyn and cs.LG | (2201.06628v4)

Abstract: In this paper, we present a deep learning technique for data-driven predictions of wave propagation in a fluid medium. The technique relies on an attention-based convolutional recurrent autoencoder network (AB-CRAN). To construct a low-dimensional representation of wave propagation data, we employ a denoising-based convolutional autoencoder. The AB-CRAN architecture with attention-based long short-term memory cells forms our deep neural network model for the time marching of the low-dimensional features. We assess the proposed AB-CRAN framework against the standard recurrent neural network for the low-dimensional learning of wave propagation. To demonstrate the effectiveness of the AB-CRAN model, we consider three benchmark problems, namely, one-dimensional linear convection, the nonlinear viscous Burgers equation, and the two-dimensional Saint-Venant shallow water system. Using the spatial-temporal datasets from the benchmark problems, our novel AB-CRAN architecture accurately captures the wave amplitude and preserves the wave characteristics of the solution for long time horizons. The attention-based sequence-to-sequence network increases the time-horizon of prediction compared to the standard recurrent neural network with long short-term memory cells. The denoising autoencoder further reduces the mean squared error of prediction and improves the generalization capability in the parameter space.

Citations (23)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.