Papers
Topics
Authors
Recent
Search
2000 character limit reached

Compositional amortized inference for large-scale hierarchical Bayesian models

Published 20 May 2025 in q-bio.QM | (2505.14429v1)

Abstract: Amortized Bayesian inference (ABI) has emerged as a powerful simulation-based approach for estimating complex mechanistic models, offering fast posterior sampling via generative neural networks. However, extending ABI to hierarchical models, a cornerstone of modern Bayesian analysis, remains a major challenge due to the difficulty of scaling to large numbers of parameters. In this work, we build on compositional score matching (CSM), a divide-and-conquer strategy for Bayesian updating using diffusion models. To address existing stability issues of CSM, we propose adaptive solvers coupled with a novel, error-damping compositional estimator. Our proposed method remains stable even with hundreds of thousands of data points and parameters. We validate our approach on a controlled toy example, a high-dimensional spatial autoregressive model, and a real-world advanced microscopy biological application task involving over 750,000 parameters.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 0 likes about this paper.