Papers
Topics
Authors
Recent
Search
2000 character limit reached

Iterated INLA for State and Parameter Estimation in Nonlinear Dynamical Systems

Published 26 Feb 2024 in stat.ML and cs.LG | (2402.17036v2)

Abstract: Data assimilation (DA) methods use priors arising from differential equations to robustly interpolate and extrapolate data. Popular techniques such as ensemble methods that handle high-dimensional, nonlinear PDE priors focus mostly on state estimation, however can have difficulty learning the parameters accurately. On the other hand, machine learning based approaches can naturally learn the state and parameters, but their applicability can be limited, or produce uncertainties that are hard to interpret. Inspired by the Integrated Nested Laplace Approximation (INLA) method in spatial statistics, we propose an alternative approach to DA based on iteratively linearising the dynamical model. This produces a Gaussian Markov random field at each iteration, enabling one to use INLA to infer the state and parameters. Our approach can be used for arbitrary nonlinear systems, while retaining interpretability, and is furthermore demonstrated to outperform existing methods on the DA task. By providing a more nuanced approach to handling nonlinear PDE priors, our methodology offers improved accuracy and robustness in predictions, especially where data sparsity is prevalent.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (36)
  1. Latent force models. In Artificial Intelligence and Statistics, pages 9–16. PMLR, 2009.
  2. Particle markov chain monte carlo methods. Journal of the Royal Statistical Society Series B: Statistical Methodology, 72(3):269–342, 2010.
  3. Optical tomography: forward and inverse problems. Inverse problems, 25(12):123010, 2009.
  4. M. Baer. findiff software package, 2018. URL https://github.com/maroba/findiff.
  5. Ross N Bannister. A review of operational methods of variational and ensemble-variational data assimilation. Quarterly Journal of the Royal Meteorological Society, 143(703):607–633, 2017.
  6. Curse-of-dimensionality revisited: Collapse of the particle filter in very large scale systems. In Probability and statistics: Essays in honor of David A. Freedman, volume 2, pages 316–335. Institute of Mathematical Statistics, 2008.
  7. Joint state and parameter estimation with an iterative ensemble Kalman smoother. Nonlinear Processes in Geophysics, 20(5):803–818, 2013.
  8. Algorithm 887: CHOLMOD, supernodal sparse Cholesky factorization and update/downdate. ACM Transactions on Mathematical Software (TOMS), 35(3):1–14, 2008.
  9. Solving and learning nonlinear PDEs with Gaussian processes. Journal of Computational Physics, 447:110668, 2021.
  10. An introduction to sequential Monte Carlo, volume 4. Springer, 2020.
  11. State and parameter estimation in stochastic dynamical models. Physica D: Nonlinear Phenomena, 239(18):1781–1788, 2010.
  12. Geir Evensen. Sequential data assimilation with a nonlinear quasi-geostrophic model using Monte Carlo methods to forecast error statistics. Journal of Geophysical Research: Oceans, 99(C5):10143–10162, 1994.
  13. Geir Evensen. The ensemble Kalman filter for combined state and parameter estimation. IEEE Control Systems Magazine, 29(3):83–104, 2009.
  14. An ensemble Kalman smoother for nonlinear dynamics. Monthly Weather Review, 128(6):1852–1867, 2000.
  15. Data assimilation fundamentals: A unified formulation of the state and parameter estimation problem. Springer Nature, 2022.
  16. A kernel two-sample test. The Journal of Machine Learning Research, 13(1):723–773, 2012.
  17. Fourth-order time-stepping for stiff PDEs. SIAM Journal on Scientific Computing, 26(4):1214–1233, 2005.
  18. Variational algorithms for analysis and assimilation of meteorological observations: theoretical aspects. Tellus A: Dynamic Meteorology and Oceanography, 38(2):97–110, 1986.
  19. Bayesian spatial modelling with R-INLA. Journal of statistical software, 63(19), 2015.
  20. Iterative linearised INLA method, 2023. URL https://inlabru-org.github.io/inlabru/articles/method.html. From the inlabru 2.10.1 package documentation.
  21. An explicit link between Gaussian fields and Gaussian Markov random fields: the stochastic partial differential equation approach. Journal of the Royal Statistical Society Series B: Statistical Methodology, 73(4):423–498, 2011.
  22. AutoIP: A united framework to integrate physics into Gaussian processes. In International Conference on Machine Learning, pages 14210–14222. PMLR, 2022.
  23. Stochastic partial differential equations. Springer, 2017.
  24. First application of data assimilation-based control to fusion plasma. Scientific Reports, 14(1):137, 2024.
  25. Non-separable spatio-temporal graph kernels via SPDEs. In International Conference on Artificial Intelligence and Statistics, pages 10640–10660. PMLR, 2022.
  26. Numerical optimization. Springer, 1999.
  27. UK Met Office. Data assimilation methods, 2024. URL https://www.metoffice.gov.uk/research/weather/satellite-and-surface-assimilation/data-assimilation-methods.
  28. nansencenter/DAPPER, 2018. URL https://doi.org/10.5281/zenodo.2029296.
  29. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. Journal of Computational physics, 378:686–707, 2019.
  30. Gaussian Processes for Machine Learning. MIT Press, 2006.
  31. R-INLA: approximate bayesian inference using integrated nested laplace approximations. Trondheim, Norway. URL http://www. r-inla. org, 23:115, 2013.
  32. Approximate Bayesian inference for hierarchical Gaussian Markov random field models. Journal of statistical planning and inference, 137(10):3177–3192, 2007.
  33. Approximate Bayesian Inference for Latent Gaussian models by using Integrated Nested Laplace Approximations. Journal of the Royal Statistical Society Series B: Statistical Methodology, 71(2):319–392, 04 2009. ISSN 1369-7412. 10.1111/j.1467-9868.2008.00700.x. URL https://academic.oup.com/jrsssb/article/71/2/319/7092907.
  34. Earth system modeling 2.0: A blueprint for models that learn from observations and targeted high-resolution simulations. Geophysical Research Letters, 44(24):12–396, 2017.
  35. Kazuhiro Takahashi. Formation of sparse bus impedance matrix and its application to short circuit study. In Proc. PICA Conference, June, 1973, 1973.
  36. Accurate approximations for posterior moments and marginal densities. Journal of the american statistical association, 81(393):82–86, 1986.
Citations (1)

Summary

  • The paper introduces an iterative INLA method that integrates state and parameter estimation via successive linearization of nonlinear models.
  • It transforms nonlinear PDE priors into Gaussian Markov Random Fields, enabling efficient application of INLA for improved inference.
  • Practical experiments demonstrate enhanced prediction accuracy and robustness compared to traditional data assimilation techniques.

Iterated INLA for Enhanced Data Assimilation in Nonlinear Dynamical Systems

Introduction

In traditional approaches to data assimilation (DA) in systems governed by nonlinear partial differential equations (PDE), challenges often arise due to the computational complexity and the difficulty in accurately estimating model parameters. Ensemble methods and machine learning-based alternatives each offer partial solutions, yet they either struggle with parameter estimation or entail interpretability issues concerning uncertainties. This paper introduces an advanced DA methodology leveraging Integrated Nested Laplace Approximation (INLA), specifically designed for arbitrary nonlinear systems. By iteratively linearizing the dynamical model to form a Gaussian Markov Random Field (GMRF), the method succinctly integrates state and parameter inference, thereby providing a significant improvement in prediction accuracy and robustness, particularly in data-sparse scenarios.

Methodology

The core of our approach revolves around the SPDE-INLA technique, extending it to suit nonlinear PDE priors via iterative linearization. This results in a direct representation of the GP priors as GMRFs, onto which INLA is applied for efficient state and parameter estimation. The method asserts two primary strengths: the ability to employ priors derived from any nonlinear PDE and to furnish accurate estimates of non-Gaussian state and posterior distributions. By also addressing model uncertainty and incorporating expert knowledge, our approach significantly enhances interpretability and predictability in DA tasks.

Theoretical Implications

Our methodology's theoretical foundation capitalizes on the elegant framework provided by INLA for latent Gaussian models. Through the iterative linearization of the nonlinear dynamical models, we construct an adaptable framework that resonates well with the inherent complexities of nonlinear PDE systems. The feasibility of extending this approach to handle non-Gaussian likelihoods and exploring nested Laplace approximations presents a promising avenue for future research. Additionally, the convergence of our algorithm to the solutions provided by the Gauß-Newton method reinforces our method's reliability and effectiveness.

Practical Implications

Practically, the implementation of our iterative INLA method showcases its utility in accurately recovering states and parameters from sparse and noisy data across various nonlinear PDE systems. The experiments conducted illustrate the method's robustness and its superior performance against established DA techniques. Especially in scenarios where traditional methods face numerical stability issues or potential overfitting due to incorrect model assumptions, our method exhibits remarkable resilience and consistency. Its application extends potential benefits to fields reliant on complex dynamical models, including environmental sciences, engineering, and applied physics.

Future Prospects

Looking ahead, the development of more scalable versions of the proposed methodology could address current limitations regarding high-dimensional problems. Integrating the method with filtering/smoothing strategies specific to temporal data may also expedite computations and enhance performance. The exploration into non-Gaussian likelihoods and more complex model dynamics, including stochastic components and multi-scale phenomena, will likely broaden the method's applicability. As the method matures, its contributions to improving model predictions, understanding system uncertainties, and guiding empirical studies could be substantial.

Conclusion

In summation, the proposed iterated INLA method provides an innovative and theoretically sound approach to the complex challenges of DA in nonlinear dynamical systems. Its ability to combine accurate state and parameter estimation with improved computational efficiency and robustness marks a significant advancement in the field. As we continue to explore its extensions and applications, the full potential of this method in contributing to various scientific and engineering disciplines remains promising and widely anticipated.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 35 likes about this paper.