Papers
Topics
Authors
Recent
Search
2000 character limit reached

An overview of differentiable particle filters for data-adaptive sequential Bayesian inference

Published 19 Feb 2023 in cs.LG and cs.AI | (2302.09639v2)

Abstract: By approximating posterior distributions with weighted samples, particle filters (PFs) provide an efficient mechanism for solving non-linear sequential state estimation problems. While the effectiveness of particle filters has been recognised in various applications, their performance relies on the knowledge of dynamic models and measurement models, as well as the construction of effective proposal distributions. An emerging trend involves constructing components of particle filters using neural networks and optimising them by gradient descent, and such data-adaptive particle filtering approaches are often called differentiable particle filters. Due to the expressiveness of neural networks, differentiable particle filters are a promising computational tool for performing inference on sequential data in complex, high-dimensional tasks, such as vision-based robot localisation. In this paper, we review recent advances in differentiable particle filters and their applications. We place special emphasis on different design choices for key components of differentiable particle filters, including dynamic models, measurement models, proposal distributions, optimisation objectives, and differentiable resampling techniques.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (101)
  1. R. E. Kalman, “A new approach to linear filtering and prediction problems,” J .Basic Eng., vol. 82, no. 1, pp. 33–45, 1960.
  2. B. Anderson and J. B. Moore, “Optimal filtering,” Prentice-Hall Inform. and Syst. Sci. Ser., 1979.
  3. F. E. Daum, “Extended Kalman filters.” Encyclo. Syst. and Control, 2015.
  4. E. A. Wan and R. Van Der Merwe, “The unscented kalman filter for nonlinear estimation,” in Proc. IEEE Adapt. Syst. Signal. Process. Commun. Control Sympos., Oct., Lake Louise, Canada 2000.
  5. R. S. Bucy and K. D. Senne, “Digital synthesis of non-linear filters,” Automatica, vol. 7, no. 3, pp. 287–298, 1971.
  6. A. Doucet, D. N. Freitas, and N. Gordon, “An introduction to sequential Monte Carlo methods,” in Sequential Monte Carlo methods in practice.   Springer, 2001, pp. 3–14.
  7. N. Gordon, D. Salmond, and A. Smith, “Novel approach to nonlinear/non-Gaussian Bayesian state estimation,” in IEE Proc. F (Radar and Signal Process.), vol. 140, 1993, pp. 107–113.
  8. P. M. Djurić, J. H. Kotecha, J. Zhang, Y. Huang, T. Ghirmai, M. F. Bugallo, and J. Miguez, “Particle filtering,” IEEE Signal Process. Mag., vol. 20, no. 5, pp. 19–38, 2003.
  9. P. Del Moral and L. Miclo, “Branching and interacting particle systems. approximations of Feynman-Kac formulae with applications to non-linear filtering,” Sémin. Probab. Strasbourg, vol. 34, pp. 1–145, 2000.
  10. D. Crisan and A. Doucet, “A survey of convergence results on particle filtering methods for practitioners,” IEEE Trans. Signal Process., vol. 50, no. 3, pp. 736–746, 2002.
  11. P. Del Moral, “Measure-valued processes and interacting particle systems. application to nonlinear filtering problems,” Ann. Appl. Probab., vol. 8, no. 2, pp. 438–495, 1998.
  12. V. Elvira, J. Míguez, and P. M. Djurić, “Adapting the number of particles in sequential Monte Carlo methods through an online scheme for convergence assessment,” IEEE Trans. Signal Process., vol. 65, no. 7, pp. 1781–1794, 2017.
  13. V. Elvira, J. Miguez, and P. M. Djurić, “On the performance of particle filters with adaptive number of particles,” Stat. Comput., vol. 31, pp. 1–18, 2021.
  14. M. K. Pitt and N. Shephard, “Filtering via simulation: Auxiliary particle filters,” J. Amer. Statist. Assoc., vol. 94, no. 446, pp. 590–599, 1999.
  15. V. Elvira, L. Martino, M. F. Bugallo, and P. M. Djurić, “Elucidating the auxiliary particle filter via multiple importance sampling,” IEEE Signal Process. Mag., vol. 36, no. 6, pp. 145–152, 2019.
  16. N. Branchini and V. Elvira, “Optimized auxiliary particle filters: adapting mixture proposals via convex optimization,” in Proc. Conf. Uncertain. Artif. Intell. (UAI), 2021, pp. 1289–1299.
  17. J. H. Kotecha and P. M. Djurić, “Gaussian sum particle filtering,” IEEE Trans. Signal Process., vol. 51, no. 10, pp. 2602–2612, 2003.
  18. ——, “Gaussian sum particle filtering for dynamic state-space models,” in Proc. IEEE Int. Conf. Acoust. Speech Signal Process., Salt Lake City, USA, May 2001.
  19. ——, “Gaussian particle filtering,” IEEE Trans. Signal Process., vol. 51, no. 10, pp. 2592–2601, 2003.
  20. R. Van Der Merwe, A. Doucet, N. De Freitas, and E. Wan, “The unscented particle filter,” Proc. Adv. Neur. Inf. Process. Sys. (NeurIPS), Dec. 2000.
  21. Y. Rui and Y. Chen, “Better proposal distributions: Object tracking using unscented particle filter,” in IEEE Conf. Comp. Vis. and Pat. Recog. (CVPR).   Kauai, USA.: IEEE, Dec. 2001.
  22. S. J. Julier and J. K. Uhlmann, “Unscented filtering and nonlinear estimation,” Proc. IEEE, vol. 92, no. 3, pp. 401–422, 2004.
  23. A. Doucet, N. de Freitas, K. Murphy, and S. Russell, “Rao-Blackwellised particle filtering for dynamic Bayesian networks,” in Proc. Conf. Uncertain. Artif. Intell. (UAI), Stanford, USA, 2000, pp. 176–183.
  24. N. De Freitas, “Rao-Blackwellised particle filtering for fault diagnosis,” in Proc. IEEE Aerosp. Conf., vol. 4, 2002, pp. 4–4.
  25. P. M. Djurić, T. Lu, and M. F. Bugallo, “Multiple particle filtering,” in Proc. IEEE Int. Conf. Acoust. Speech Signal Process. (ICASSP), Honolulu, USA, Apr. 2007.
  26. P. M. Djurić and M. F. Bugallo, “Particle filtering for high-dimensional systems,” in Proc. Comput. Adv. Multi-Sensor Adapt. Process. (CAMSAP), Saint Martin, France, Dec. 2013.
  27. P. Bunch and S. Godsill, “Approximations of the optimal importance density using gaussian particle flow importance sampling,” J. Amer. Statist. Assoc., vol. 111, no. 514, pp. 748–762, 2016.
  28. Y. Li and M. Coates, “Particle filtering with invertible particle flow,” IEEE Trans. Signal Process., vol. 65, no. 15, pp. 4102–4116, 2017.
  29. J. Heng, A. Doucet, and Y. Pokern, “Gibbs flow for approximate transport with applications to Bayesian computation,” J. R. Stat. Soc. Ser. B. Stat. Methodol, vol. 83, no. 1, pp. 156–187, 2021.
  30. T. Zhang, C. Xu, and M.-H. Yang, “Multi-task correlation particle filter for robust object tracking,” in Proc. IEEE Conf. Comput. Vis. and Pattern Recogn. (CVPR), Honolulu, Hawaii, July 2017.
  31. S. Malik and M. K. Pitt, “Particle filters for continuous likelihood evaluation and maximisation,” J. Econometrics, vol. 165, no. 2, pp. 190–209, 2011.
  32. A. Gunatilake, S. Kodagoda, and K. Thiyagarajan, “A novel UHF-RFID dual antenna signals combined with Gaussian process and particle filter for in-pipe robot localization,” IEEE Robot. Autom. Lett., vol. 7, no. 3, pp. 6005–6011, 2022.
  33. P. J. Van Leeuwen, H. R. Künsch, L. Nerger, R. Potthast, and S. Reich, “Particle filters for high-dimensional geoscience applications: A review,” Q. J. R. Meteorol. Soc., vol. 145, no. 723, pp. 2335–2365, 2019.
  34. C. Pozna, R.-E. Precup, E. Horváth, and E. M. Petriu, “Hybrid particle filter–particle swarm optimization algorithm and application to fuzzy controlled servo systems,” IEEE Trans. Fuzzy Syst., vol. 30, no. 10, pp. 4286–4297, 2022.
  35. N. Kantas, A. Doucet, S. S. Singh, and J. M. Maciejowski, “An overview of sequential Monte Carlo methods for parameter estimation in general state-space models,” IFAC Proc. Vol., vol. 42, no. 10, pp. 774–785, 2009.
  36. N. Kantas, A. Doucet, S. S. Singh, J. Maciejowski, and N. Chopin, “On particle methods for parameter estimation in state-space models,” Stat. Sci., vol. 30, no. 3, pp. 328–351, 2015.
  37. C. Andrieu, A. Doucet, and R. Holenstein, “Particle Markov chain Monte Carlo methods,” J. R. Stat. Soc. Ser. B. Stat. Methodol., vol. 72, no. 3, pp. 269–342, 2010.
  38. F. Lindsten, M. I. Jordan, and T. B. Schon, “Particle gibbs with ancestor sampling,” J. Mach. Learn. Res., vol. 15, pp. 2145–2184, 2014.
  39. N. Chopin, P. E. Jacob, and O. Papaspiliopoulos, “SMC2: an efficient algorithm for sequential analysis of state space models,” J. R. Stat. Soc. Ser. B. Stat. Methodol., vol. 75, no. 3, pp. 397–426, 2013.
  40. S. Pérez-Vieites, I. P. Mariño, and J. Míguez, “Probabilistic scheme for joint parameter estimation and state prediction in complex dynamical systems,” Phys. Rev. E, vol. 98, no. 6, p. 063305, 2018.
  41. D. Crisan and J. MÍguez, “Nested particle filters for online parameter estimation in discrete-time state-space Markov models,” Bernoulli, vol. 24, no. 4A, pp. 3039–3086, 2018.
  42. S. Pérez-Vieites and J. Míguez, “Nested Gaussian filters for recursive Bayesian inference and nonlinear tracking in state space models,” Signal Proces., vol. 189, p. 108295, 2021.
  43. E. Chouzenoux and V. Elvira, “GraphEM: EM algorithm for blind Kalman filtering under graphical sparsity constraints,” in Proc. IEEE Int. Conf. Acoust. Speech Signal Process. (ICASSP), May 2020.
  44. V. Elvira and É. Chouzenoux, “Graphical inference in linear-Gaussian state-space models,” IEEE Trans. Signal Process., vol. 70, pp. 4757–4771, 2022.
  45. E. Chouzenoux and V. Elvira, “Sparse graphical linear dynamical systems,” arXiv preprint arXiv:2307.03210, 2023.
  46. R. Jonschkowski, D. Rastogi, and O. Brock, “Differentiable particle filters: end-to-end learning with algorithmic priors,” in Proc. Robot.: Sci. and Syst. (RSS), Pittsburgh, Pennsylvania, July 2018.
  47. P. Karkus, D. Hsu, and W. S. Lee, “Particle filter networks with application to visual localization,” in Proc. Conf. Robot Learn. (CoRL), Zurich, Switzerland, Oct 2018.
  48. H. Wen, X. Chen, G. Papagiannis, C. Hu, and Y. Li, “End-to-end semi-supervised learning for differentiable particle filters,” in Proc. IEEE Int. Conf. Robot. Automat., (ICRA), Xi’an, China, May 2021.
  49. X. Chen, H. Wen, and Y. Li, “Differentiable particle filters through conditional normalizing flow,” in Proc. IEEE Int. Conf. Inf. Fusion (FUSION), Sun City, South Africa, Nov. 2021.
  50. X. Chen and Y. Li, “Conditional measurement density estimation in sequential Monte Carlo via normalizing flow,” in Proc. Euro. Sig. Process. Conf. (EUSIPCO), Belgrade, Serbia, Aug. 2022.
  51. A. Corenflos, J. Thornton, G. Deligiannidis, and A. Doucet, “Differentiable particle filtering via entropy-regularized optimal transport,” in Proc. Int. Conf. Mach. Learn. (ICML), July 2021.
  52. M. Zhu, K. Murphy, and R. Jonschkowski, “Towards differentiable resampling,” arXiv preprint arXiv:2004.11938, 2020.
  53. V. Elvira and L. Martino, “Advances in importance sampling,” Wiley StatsRef-Statistics Reference Online, pp. 1–14, 2021.
  54. T. Hesterberg, “Weighted average importance sampling and defensive mixture distributions,” Technometrics, vol. 37, no. 2, pp. 185–194, 1995.
  55. S. T. Tokdar and R. E. Kass, “Importance sampling: a review,” Wiley Interdiscip. Rev. Comput. Stat., vol. 2, no. 1, pp. 54–60, 2010.
  56. S. Agapiou, O. Papaspiliopoulos, D. Sanz-Alonso, and A. M. Stuart, “Importance sampling: Intrinsic dimension and computational cost,” Statist. Sci., pp. 405–431, 2017.
  57. P. Bickel, B. Li, and T. Bengtsson, “Sharp failure rates for the bootstrap particle filter in high dimensions,” Pushing the limits of contemporary statistics: Contributions in honor of Jayanta K. Ghosh, 2008.
  58. T. Li, M. Bolic, and P. M. Djuric, “Resampling methods for particle filtering: classification, implementation, and strategies,” IEEE Signal Process. Mag., vol. 32, no. 3, pp. 70–86, 2015.
  59. M. Gerber, N. Chopin, and N. Whiteley, “Negative association, ordering and convergence of resampling methods,” Ann. Statist., vol. 47, no. 4, pp. 2236–2260, 2019.
  60. T.-c. Li, G. Villarrubia, S.-d. Sun, J. M. Corchado, and J. Bajo, “Resampling methods for particle filtering: identical distribution, a new method, and comparable study,” Frontiers Inform. Tech. & Electron. Eng., vol. 16, no. 11, pp. 969–984, 2015.
  61. R. Douc and O. Cappé, “Comparison of resampling schemes for particle filtering,” in Proc. Int. Symp. Image and Signal Process. and Anal., Zagreb, Croatia, 2005.
  62. M. Bolić, P. M. Djurić, and S. Hong, “Resampling algorithms for particle filters: A computational complexity perspective,” EURASIP J. Adv. Signal Process., vol. 2004, no. 15, pp. 1–11, 2004.
  63. A. Doucet, S. Godsill, and C. Andrieu, “On sequential Monte Carlo sampling methods for bayesian filtering,” Stat. Comput., vol. 10, no. 3, pp. 197–208, 2000.
  64. V. Elvira, L. Martino, and C. P. Robert, “Rethinking the effective sample size,” Int. Stat. Rev., vol. 90, no. 3, pp. 525–550, 2022.
  65. L. Martino, V. Elvira, and F. Louzada, “Effective sample size for importance sampling based on discrepancy measures,” Signal Process., vol. 131, pp. 386–401, 2017.
  66. J. H. Huggins and D. M. Roy, “Convergence of sequential Monte Carlo based sampling methods,” arXiv preprint arXiv:1503.00966, 2015.
  67. Z. Li, P. Fan, and Y. Dong, “Flexible effective sample size based on the message importance measure,” IEEE Open J. Signal Process., vol. 1, pp. 216–229, 2020.
  68. A. Doucet and V. B. Tadić, “Parameter estimation in general state-space models using particle methods,” Ann. Inst. Stat. Math., vol. 55, no. 2, pp. 409–422, 2003.
  69. C. Andrieu, A. Doucet, and V. B. Tadic, “On-line parameter estimation in general state-space models,” in Proc. IEEE Conf. Dec. and Contr. (CDC), Seville, Spain, Dec. 2005.
  70. A. Kloss, G. Martius, and J. Bohg, “How to train your differentiable filter,” Auto. Robot., vol. 45, no. 4, pp. 561–578, 2021.
  71. C. Rosato, L. Devlin, V. Beraud, P. Horridge, T. B. Schön, and S. Maskell, “Efficient learning of the parameters of non-linear models using differentiable resampling in particle filters,” IEEE Trans. Signal Process., vol. 70, pp. 3676–3692, 2022.
  72. R. J. Williams, “Simple statistical gradient-following algorithms for connectionist reinforcement learning,” Mach. Learn., vol. 8, no. 3-4, pp. 229–256, 1992.
  73. D. P. Kingma and M. Welling, “Auto-encoding variational Bayes,” in Proc. Int. Conf. Learn. Represent. (ICLR), Scottsdale, Arizona, May 2013.
  74. S. Reich, “A nonparametric ensemble transform method for Bayesian inference,” SIAM J. Sci. Comput., vol. 35, no. 4, pp. A2013–A2024, 2013.
  75. M. Cuturi, “Sinkhorn distances: Lightspeed computation of optimal transport,” in Proc. Adv. Neur. Inf. Process. Sys. (NeurIPS), Lake Tahoe, USA, Dec. 2013.
  76. J. Feydy et al., “Interpolating between optimal transport and mmd using Sinkhorn divergences,” in Proc. Int. Conf. Artif. Intell. Stat. (AISTAS), Naha, Japan, Apr. 2019.
  77. G. Peyré and M. Cuturi, “Computational optimal transport,” Foundations and Trends® in Machine Learning, vol. 11, no. 5-6, pp. 355–607, 2019.
  78. J. Lee et al., “Set transformer: A framework for attention-based permutation-invariant neural networks,” in Proc. Int. Conf. Mach. Learn. (ICML), Baltimore, USA, June 2019.
  79. A. Vaswani et al., “Attention is all you need,” in Proc. Adv. Neur. Inf. Process. Sys. (NeurIPS), Long Beach, USA, Dec. 2017.
  80. X. Ma, P. Karkus, D. Hsu, and W. S. Lee, “Particle filter recurrent neural networks,” in Proc. AAAI Conf. Artif. Intell. (AAAI), New York, USA, Feb. 2020.
  81. X. Ma et al., “Discriminative particle filter reinforcement learning for complex partial observations,” in Proc. Int. Conf. Learn. Rep. (ICLR), New Orleans, USA, May 2019.
  82. R. Chen, H. Yin, Y. Jiao, G. Dissanayake, Y. Wang, and R. Xiong, “Deep samplable observation model for global localization and kidnapping,” IEEE Robot. Autom. Lett. (RAL), vol. 6, no. 2, pp. 2296–2303, 2021.
  83. P. Karkus, S. Cai, and D. Hsu, “Differentiable slam-net: Learning particle slam for visual navigation,” in Proc. IEEE Conf. Comp. Vis. and Pat. Recog. (CVPR), June 2021.
  84. P. Karkus et al., “Differentiable algorithm networks for composable robot learning,” in Proc. Robot.: Sci. and Syst. (RSS), Messe Freiburg, Germany, June 2019.
  85. M. A. Lee, B. Yi, R. Martín-Martín, S. Savarese, and J. Bohg, “Multimodal sensor fusion with differentiable filters,” in Proc. IEEE/RSJ Int. Conf. Intel. Robot. Sys. (IROS), Las Vegas, USA, Oct. 2020.
  86. C. Naesseth, S. Linderman, R. Ranganath, and D. Blei, “Variational sequential Monte Carlo,” in Proc. Int. Conf. Artif. Intel. and Stat. (AISTATS), Playa Blanca, Spain, Apr. 2018.
  87. T. A. Le, M. Igl, T. Rainforth, T. Jin, and F. Wood, “Auto-encoding sequential Monte Carlo,” in Proc. Int. Conf. Learn. Rep. (ICLR), Vancouver, Canada, Apr. 2018.
  88. C. J. Maddison et al., “Filtering variational objectives,” in Proc. Adv. Neur. Inf. Process. Sys. (NeurIPS), Long Beach, USA, Dec. 2017.
  89. M. H. Dupty, Y. Dong, and W. S. Lee, “PF-GNN: Differentiable particle filtering based approximation of universal graph representations,” in Proc. Int. Conf. Learn. Represent. (ICLR), May 2021.
  90. G. Papamakarios, E. Nalisnick, D. J. Rezende, S. Mohamed, and B. Lakshminarayanan, “Normalizing flows for probabilistic modeling and inference,” J. Mach. Learn. Res., vol. 22, pp. 1–64, 2021.
  91. S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Comput., vol. 9, no. 8, pp. 1735–1780, 1997.
  92. K. Cho, B. van Merriënboer, D. Bahdanau, and Y. Bengio, “On the properties of neural machine translation: Encoder–decoder approaches,” Syntax, Semant. and Struct. in Stat. Transl., p. 103, 2014.
  93. F. Gama, N. Zilberstein, R. G. Baraniuk, and S. Segarra, “Unrolling particles: Unsupervised learning of sampling distributions,” in Proc. IEEE Int. Conf. Acoust. Speech Signal Process. (ICASSP), Singapore, May 2022, pp. 5498–5502.
  94. F. Gama, N. Zilberstein, M. Sevilla, R. Baraniuk, and S. Segarra, “Unsupervised learning of sampling distributions for particle filters,” arXiv preprint arXiv:2302.01174, 2023.
  95. C. Winkler, D. Worrall, E. Hoogeboom, and M. Welling, “Learning likelihoods with conditional normalizing flows,” arXiv preprint arXiv:1912.00042, 2019.
  96. M. Jaderberg, K. Simonyan, A. Zisserman, and K. Kavukcuoglu, “Spatial transformer networks,” in Proc. Adv. Neur. Inf. Process. Sys. (NeurIPS), Montreal, Canada, Dec. 2015.
  97. C. Andrieu, A. Doucet, S. Singh, and V. Tadic, “Particle methods for change detection, system identification, and control,” Proc. IEEE, vol. 92, no. 3, pp. 423–438, 2004.
  98. D. Rezende and S. Mohamed, “Variational inference with normalizing flows,” in Proc. Int. Conf. Mach. Learn. (ICML), Lille, France, July 2015.
  99. Y. Burda, R. B. Grosse, and R. Salakhutdinov, “Importance weighted autoencoders,” in Proc. Int. Conf. Learn. Rep. (ICLR), San Juan, Puerto Rico, May. 2016.
  100. M. I. Jordan, Z. Ghahramani, T. S. Jaakkola, and L. K. Saul, “An introduction to variational methods for graphical models,” Mach. Learn., vol. 37, no. 2, pp. 183–233, 1999.
  101. D. J. Rezende, S. Mohamed, and D. Wierstra, “Stochastic backpropagation and approximate inference in deep generative models,” in Proc. Int. Conf. Mach. Learn. (ICML), Beijing, China, June 2014.
Citations (14)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.