Papers
Topics
Authors
Recent
Search
2000 character limit reached

Particle Guidance: non-I.I.D. Diverse Sampling with Diffusion Models

Published 19 Oct 2023 in cs.LG and cs.AI | (2310.13102v2)

Abstract: In light of the widespread success of generative models, a significant amount of research has gone into speeding up their sampling time. However, generative models are often sampled multiple times to obtain a diverse set incurring a cost that is orthogonal to sampling time. We tackle the question of how to improve diversity and sample efficiency by moving beyond the common assumption of independent samples. We propose particle guidance, an extension of diffusion-based generative sampling where a joint-particle time-evolving potential enforces diversity. We analyze theoretically the joint distribution that particle guidance generates, how to learn a potential that achieves optimal diversity, and the connections with methods in other disciplines. Empirically, we test the framework both in the setting of conditional image generation, where we are able to increase diversity without affecting quality, and molecular conformer generation, where we reduce the state-of-the-art median error by 13% on average.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Brian DO Anderson. Reverse-time diffusion equation models. Stochastic Processes and their Applications, 1982.
  2. Geom, energy-annotated molecular conformations for property prediction and molecular generation. Scientific Data, 2022.
  3. Well-tempered metadynamics: a smoothly converging and tunable free-energy method. Physical review letters, 2008.
  4. From denoising diffusions to denoising markov models. arXiv preprint arXiv:2211.03595, 2022.
  5. Emerging properties in self-supervised vision transformers. In Proceedings of the IEEE/CVF international conference on computer vision, 2021.
  6. Kernel stein generative modeling. arXiv preprint arXiv:2007.03074, 2020.
  7. Sampling is as easy as learning the score: theory for diffusion models with minimal data assumptions. ArXiv, abs/2209.11215, 2022. URL https://api.semanticscholar.org/CorpusID:252438904.
  8. Optimal dimension dependence of the metropolis-adjusted langevin algorithm. ArXiv, abs/2012.12810, 2020.
  9. Gabriele Corso. Modeling molecular structures with intrinsic diffusion models. arXiv preprint arXiv:2302.12255, 2023.
  10. Diffdock: Diffusion steps, twists, and turns for molecular docking. arXiv preprint arXiv:2210.01776, 2022.
  11. Imre Csiszár. I-divergence geometry of probability distributions and minimization problems. The annals of probability, pp.  146–158, 1975.
  12. Annealed stein variational gradient descent. arXiv preprint arXiv:2101.09815, 2021.
  13. Riemannian score-based generative modeling. arXiv preprint arXiv:2202.02763, 2022.
  14. Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 2021.
  15. Reduce, reuse, recycle: Compositional generation with energy-based diffusion models and mcmc. In International Conference on Machine Learning. PMLR, 2023.
  16. Geomol: Torsional geometric generation of molecular 3d conformer ensembles. Advances in Neural Information Processing Systems, 2021.
  17. John Geweke. Antithetic acceleration of monte carlo integration in bayesian inference. Journal of Econometrics, 38(1-2):73–89, 1988.
  18. Conformer generation with omega: learning from the data set and the analysis of failures. Journal of chemical information and modeling, 2012.
  19. Clipscore: A reference-free evaluation metric for image captioning. In Conference on Empirical Methods in Natural Language Processing, 2021.
  20. Jonathan Ho. Classifier-free diffusion guidance. ArXiv, abs/2207.12598, 2022.
  21. Bayesian ensemble refinement by replica simulations and reweighting. The Journal of chemical physics, 2015.
  22. Illuminating protein space with a programmable generative model. BioRxiv, 2022.
  23. Torsional diffusion for molecular conformer generation. arXiv preprint arXiv:2206.01729, 2022.
  24. Brownian motion and stochastic calculus. Springer Science & Business Media, 1991.
  25. Elucidating the design space of diffusion-based generative models. Advances in Neural Information Processing Systems, 2022.
  26. Determinantal point processes for machine learning. Foundations and Trends® in Machine Learning, 5(2–3):123–286, 2012.
  27. Escaping free-energy minima. Proceedings of the national academy of sciences, 2002.
  28. Greg Landrum et al. Rdkit: A software suite for cheminformatics, computational chemistry, and predictive modeling, 2013.
  29. Microsoft coco: Common objects in context. In European Conference on Computer Vision, 2014.
  30. Stein variational gradient descent: A general purpose bayesian inference algorithm. Advances in neural information processing systems, 29, 2016.
  31. Vendi sampling for molecular simulations: Diversity as a force for faster convergence and better exploration. The Journal of Chemical Physics, 159(14), 2023.
  32. Automated exploration of the low-energy chemical space with fast quantum chemical methods. Physical Chemistry Chemical Physics, 2020.
  33. High-resolution image synthesis with latent diffusion models. 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021.
  34. Photorealistic text-to-image diffusion models with deep language understanding. ArXiv, abs/2205.11487, 2022.
  35. Laion-5b: An open large-scale dataset for training next generation image-text models. ArXiv, abs/2210.08402, 2022.
  36. Learning gradient fields for molecular conformation generation. In International Conference on Machine Learning, 2021.
  37. Richard Sinkhorn. A relationship between arbitrary positive matrices and doubly stochastic matrices. Annals of Mathematical Statistics, 35:876–879, 1964. URL https://api.semanticscholar.org/CorpusID:120846714.
  38. Diffusion art or digital forgery? investigating data replication in diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023.
  39. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021.
  40. LAION-AI Team. Laion-aesthetics predictor v2. https://github.com/christophschuhmann/improved-aesthetic-predictor, 2022.
  41. Energy-based models for sparse overcomplete representations. Journal of Machine Learning Research, 2003.
  42. Autodock vina: improving the speed and accuracy of docking with a new scoring function, efficient optimization, and multithreading. Journal of computational chemistry, 2010.
  43. Blindness of score-based methods to isolated components and mixing proportions. 2020.
  44. Geodiff: A geometric diffusion model for molecular conformation generation. In International Conference on Learning Representations, 2021.
  45. Poisson flow generative models. Advances in Neural Information Processing Systems, 2022.
  46. Restart sampling for improving generative processes. arXiv preprint arXiv:2306.14878, 2023a.
  47. Pfgm++: Unlocking the potential of physics-inspired generative models. In International Conference on Machine Learning, 2023b.
  48. Se(3) diffusion model with application to protein backbone generation. arXiv preprint, 2023.
  49. Stochastic particle-optimization sampling and the non-asymptotic convergence theory. In International Conference on Artificial Intelligence and Statistics. PMLR, 2020.
  50. Message passing stein variational gradient descent. In International Conference on Machine Learning. PMLR, 2018.
Citations (12)

Summary

  • The paper presents Particle Guidance, a technique that integrates a joint-particle potential into reverse diffusion to boost sample diversity.
  • It offers a rigorous theoretical analysis connecting the joint potential to optimal diversity, drawing insights from molecular simulations and enhanced sampling methods.
  • Empirical demonstrations in text-to-image and molecular conformer generation reveal significant improvements in output diversity and reduced median errors.

Particle Guidance: non-I.I.D. Diverse Sampling with Diffusion Models

The paper presents an innovative approach to enhance the diversity and efficiency of sample generation in diffusion models—termed "Particle Guidance" (PG). This technique advances past the common independence assumptions of generative samples, aiming to optimize diversity by introducing dependencies among them.

Key Contributions

The authors introduce a theoretically motivated framework extending diffusion-based generative sampling called particle guidance. This framework uses a joint potential across the sampling process to enforce diversity, thereby improving sample efficiency. Key contributions in the paper include:

  1. Framework Definition: Introducing the concept of particle guidance, which incorporates a joint-particle time-evolving potential in reverse diffusion processes for enhanced sample diversity.
  2. Theoretical Analysis: Providing analytic expressions for the joint potential that can achieve optimal diversity and analyzing its connection to other disciplines such as molecular simulations and enhanced sampling techniques.
  3. Empirical Demonstration: Validating particle guidance in both conditional image generation and molecular conformer generation, showing significant improvements. For instance, the method increased the diversity in the text-to-image generation and reduced median errors in molecular conformer generation.

Theoretical Framework

The paper presents a sophisticated view of how diffusion models can be extended for sampling processes. Traditional models like Stein Variational Gradient Descent are leveraged, but they face limitations in efficiently sampling complex distributions like images. Particle Guidance tackles these challenges, aiming for diversity without compromising the quality or increasing the sampling time of traditional generative models.

The joint distribution generated through particle guidance allows for the integration of additional forces such as electrostatic interactions, highlighting its utility in physically-motivated systems. The model handles time-evolving potentials in a manner that respects the underlying data distribution while moving towards a desired representation, thus offering a form of controlled sampling diversity.

Applications and Results

The empirical evaluation exhibits the potential of particle guidance in real-world applications:

  • Text-to-Image Generation: Particle guidance improved the diversity metrics in Stable Diffusion outputs, maintaining visual quality while reducing mode collapse. The method was tested on a popular data set where it showcased improvements in visual distinctness over traditional I.I.D. sampling.
  • Molecular Conformer Generation: The model outperforms current state-of-the-art methods in precision and coverage for conformer generation. Specifically, particle guidance reduced the median error versus previous models, evidencing enhanced sample spread across the distribution.

Implications and Future Work

The proposed particle guidance framework opens new pathways for sampling diverse sets efficiently. The potential applications extend beyond image generation and computational chemistry, touching AI's broader contact points with simulation disciplines. This approach may fundamentally redefining generative model efficiencies across diverse computational spectrums.

Moving forward, exploring adaptive potentials in real-time inference offers promising avenues. Additionally, theoretical exploration into the implications of modified marginal distributions by such sampling frameworks poses significant interest. The blend of machine learning with physically inspired constraints or objectives also hints at new intersections ripe for exploration.

The paper represents an integration effort across finely-grained computational methodologies and state-the-art generative techniques. While producing meaningful strides in balancing diversity and fidelity, it lays firm groundwork for deeper investigations into sample dependency generation.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 3 tweets with 18 likes about this paper.