Particle Guidance: non-I.I.D. Diverse Sampling with Diffusion Models
Abstract: In light of the widespread success of generative models, a significant amount of research has gone into speeding up their sampling time. However, generative models are often sampled multiple times to obtain a diverse set incurring a cost that is orthogonal to sampling time. We tackle the question of how to improve diversity and sample efficiency by moving beyond the common assumption of independent samples. We propose particle guidance, an extension of diffusion-based generative sampling where a joint-particle time-evolving potential enforces diversity. We analyze theoretically the joint distribution that particle guidance generates, how to learn a potential that achieves optimal diversity, and the connections with methods in other disciplines. Empirically, we test the framework both in the setting of conditional image generation, where we are able to increase diversity without affecting quality, and molecular conformer generation, where we reduce the state-of-the-art median error by 13% on average.
- Brian DO Anderson. Reverse-time diffusion equation models. Stochastic Processes and their Applications, 1982.
- Geom, energy-annotated molecular conformations for property prediction and molecular generation. Scientific Data, 2022.
- Well-tempered metadynamics: a smoothly converging and tunable free-energy method. Physical review letters, 2008.
- From denoising diffusions to denoising markov models. arXiv preprint arXiv:2211.03595, 2022.
- Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, 2021.
- Kernel stein generative modeling. arXiv preprint arXiv:2007.03074, 2020.
- Sampling is as easy as learning the score: theory for diffusion models with minimal data assumptions. ArXiv, abs/2209.11215, 2022. URL https://api.semanticscholar.org/CorpusID:252438904.
- Optimal dimension dependence of the metropolis-adjusted langevin algorithm. ArXiv, abs/2012.12810, 2020.
- Gabriele Corso. Modeling molecular structures with intrinsic diffusion models. arXiv preprint arXiv:2302.12255, 2023.
- Diffdock: Diffusion steps, twists, and turns for molecular docking. arXiv preprint arXiv:2210.01776, 2022.
- Imre Csiszár. I-divergence geometry of probability distributions and minimization problems. The annals of probability, pp. 146–158, 1975.
- Annealed stein variational gradient descent. arXiv preprint arXiv:2101.09815, 2021.
- Riemannian score-based generative modeling. arXiv preprint arXiv:2202.02763, 2022.
- Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 2021.
- Reduce, reuse, recycle: Compositional generation with energy-based diffusion models and mcmc. In International Conference on Machine Learning. PMLR, 2023.
- Geomol: Torsional geometric generation of molecular 3d conformer ensembles. Advances in Neural Information Processing Systems, 2021.
- John Geweke. Antithetic acceleration of monte carlo integration in bayesian inference. Journal of Econometrics, 38(1-2):73–89, 1988.
- Conformer generation with omega: learning from the data set and the analysis of failures. Journal of chemical information and modeling, 2012.
- Clipscore: A reference-free evaluation metric for image captioning. In Conference on Empirical Methods in Natural Language Processing, 2021.
- Jonathan Ho. Classifier-free diffusion guidance. ArXiv, abs/2207.12598, 2022.
- Bayesian ensemble refinement by replica simulations and reweighting. The Journal of chemical physics, 2015.
- Illuminating protein space with a programmable generative model. BioRxiv, 2022.
- Torsional diffusion for molecular conformer generation. arXiv preprint arXiv:2206.01729, 2022.
- Brownian motion and stochastic calculus. Springer Science & Business Media, 1991.
- Elucidating the design space of diffusion-based generative models. Advances in Neural Information Processing Systems, 2022.
- Determinantal point processes for machine learning. Foundations and Trends® in Machine Learning, 5(2–3):123–286, 2012.
- Escaping free-energy minima. Proceedings of the national academy of sciences, 2002.
- Greg Landrum et al. Rdkit: A software suite for cheminformatics, computational chemistry, and predictive modeling, 2013.
- Microsoft coco: Common objects in context. In European Conference on Computer Vision, 2014.
- Stein variational gradient descent: A general purpose bayesian inference algorithm. Advances in neural information processing systems, 29, 2016.
- Vendi sampling for molecular simulations: Diversity as a force for faster convergence and better exploration. The Journal of Chemical Physics, 159(14), 2023.
- Automated exploration of the low-energy chemical space with fast quantum chemical methods. Physical Chemistry Chemical Physics, 2020.
- High-resolution image synthesis with latent diffusion models. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021.
- Photorealistic text-to-image diffusion models with deep language understanding. ArXiv, abs/2205.11487, 2022.
- Laion-5b: An open large-scale dataset for training next generation image-text models. ArXiv, abs/2210.08402, 2022.
- Learning gradient fields for molecular conformation generation. In International Conference on Machine Learning, 2021.
- Richard Sinkhorn. A relationship between arbitrary positive matrices and doubly stochastic matrices. Annals of Mathematical Statistics, 35:876–879, 1964. URL https://api.semanticscholar.org/CorpusID:120846714.
- Diffusion art or digital forgery? investigating data replication in diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023.
- Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021.
- LAION-AI Team. Laion-aesthetics predictor v2. https://github.com/christophschuhmann/improved-aesthetic-predictor, 2022.
- Energy-based models for sparse overcomplete representations. Journal of Machine Learning Research, 2003.
- Autodock vina: improving the speed and accuracy of docking with a new scoring function, efficient optimization, and multithreading. Journal of computational chemistry, 2010.
- Blindness of score-based methods to isolated components and mixing proportions. 2020.
- Geodiff: A geometric diffusion model for molecular conformation generation. In International Conference on Learning Representations, 2021.
- Poisson flow generative models. Advances in Neural Information Processing Systems, 2022.
- Restart sampling for improving generative processes. arXiv preprint arXiv:2306.14878, 2023a.
- Pfgm++: Unlocking the potential of physics-inspired generative models. In International Conference on Machine Learning, 2023b.
- Se(3) diffusion model with application to protein backbone generation. arXiv preprint, 2023.
- Stochastic particle-optimization sampling and the non-asymptotic convergence theory. In International Conference on Artificial Intelligence and Statistics. PMLR, 2020.
- Message passing stein variational gradient descent. In International Conference on Machine Learning. PMLR, 2018.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.