Papers
Topics
Authors
Recent
Search
2000 character limit reached

Accelerated Bayesian inference using deep learning

Published 26 Mar 2019 in astro-ph.CO and astro-ph.IM | (1903.10860v1)

Abstract: We present a novel Bayesian inference tool that uses a neural network to parameterise efficient Markov Chain Monte-Carlo (MCMC) proposals. The target distribution is first transformed into a diagonal, unit variance Gaussian by a series of non-linear, invertible, and non-volume preserving flows. Neural networks are extremely expressive, and can transform complex targets to a simple latent representation. Efficient proposals can then be made in this space, and we demonstrate a high degree of mixing on several challenging distributions. Parameter space can naturally be split into a block diagonal speed hierarchy, allowing for fast exploration of subspaces where it is inexpensive to evaluate the likelihood. Using this method, we develop a nested MCMC sampler to perform Bayesian inference and model comparison, finding excellent performance on highly curved and multi-modal analytic likelihoods. We also test it on {\em Planck} 2015 data, showing accurate parameter constraints, and calculate the evidence for simple one-parameter extensions to LCDM in $\sim20$ dimensional parameter space. Our method has wide applicability to a range of problems in astronomy and cosmology.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (1)

Collections

Sign up for free to add this paper to one or more collections.