Papers
Topics
Authors
Recent
Search
2000 character limit reached

Interleaved Gibbs Diffusion: Generating Discrete-Continuous Data with Implicit Constraints

Published 19 Feb 2025 in cs.LG and cs.AI | (2502.13450v2)

Abstract: We introduce Interleaved Gibbs Diffusion (IGD), a novel generative modeling framework for discrete-continuous data, focusing on problems with important, implicit and unspecified constraints in the data. Most prior works on discrete and discrete-continuous diffusion assume a factorized denoising distribution, which can hinder the modeling of strong dependencies between random variables in such problems. We empirically demonstrate a significant improvement in 3-SAT performance out of the box by switching to a Gibbs-sampling style discrete diffusion model which does not assume factorizability. Motivated by this, we introduce IGD which generalizes discrete time Gibbs sampling type Markov chain for the case of discrete-continuous generation. IGD allows for seamless integration between discrete and continuous denoisers while theoretically guaranteeing exact reversal of a suitable forward process. Further, it provides flexibility in the choice of denoisers, allows conditional generation via state-space doubling and inference time refinement. Empirical evaluations on three challenging generation tasks - molecule structures, layouts and tabular data - demonstrate state-of-the-art performance. Notably, IGD achieves state-of-the-art results without relying on domain-specific inductive biases like equivariant diffusion or auxiliary losses. We explore a wide range of modeling, and interleaving strategies along with hyperparameters in each of these problems.

Summary

  • The paper introduces Interleaved Gibbs Diffusion, which interleaves discrete and continuous denoising in a Gibbs sampling framework to model constrained, mixed data without needing distribution factorization.
  • It incorporates the ReDeNoise algorithm and state-space doubling to enhance conditional generation and inference efficiency in complex domains like layout and molecule generation.
  • Empirical evaluations demonstrate a 7% improvement in 3-SAT task performance along with robust results in layout and molecular generation, underscoring its state-of-the-art capability.

"Interleaved Gibbs Diffusion: Generating Discrete-Continuous Data with Implicit Constraints" (2502.13450)

Introduction

The paper presents Interleaved Gibbs Diffusion (IGD), an innovative generative modeling framework designed for the effective handling of mixed continuous-discrete data within constrained generation tasks. While traditional diffusion models have demonstrated proficiency in discrete and continuous settings through factorized denoising distributions, IGD addresses a critical limitation in modeling strong dependencies among random variables in constrained environments. By interleaving continuous and discrete denoising algorithms through a Gibbs sampling-like Markov chain, IGD circumvents the need for distribution factorizability. The versatility of the IGD framework enables the incorporation of various denoisers, supports conditional generation via state-space doubling, and improves inference time efficiency through the ReDeNoise method. Empirical results from complex domains like 3-SAT, molecule generation, and layout generation underscore IGD's state-of-the-art performance, highlighting a 7% improvement in 3-SAT tasks compared to established models.

Model Framework

Interleaved Gibbs Diffusion

IGD employs a novel strategy for constrained generative tasks by interleaving discrete and continuous denoising processes. It uses a Gibbs sampling-inspired approach to denoise mixed sequences of continuous vectors and discrete tokens one element at a time, providing flexibility in choosing denoisers without requiring the factorization of the denoising process.

  • Forward Noising Process: The noising is performed through a sequence that independently applies either discrete or continuous noise at each position. Discrete noising uses a noise schedule applied through additional token spaces, while continuous noising utilizes a monotonically increasing noise schedule. This strategic separation allows for direct sampling of noised states without sequential dependency, facilitating efficient learning.
  • Reverse Denoising Process: The reverse process depends on accurately trained discrete and continuous denoisers, non-factorizable but capable of exact reversal of the forward process. Continuous denoising relies on adapted Tweedie's formula, enabling restoration of continuous elements from sequences with changing conditioning.

Enhancements in the Framework

ReDeNoise Algorithm: An additional inference strategy, ReDeNoise utilizes iterative noise-denoise cycles to enhance generation accuracy at the cost of increased computational resources, demonstrating potential accuracy gains in practical applications.

Conditional Generation: The framework adapts state-space doubling, a strategy inspired by Domain Layout Transformer (DLT), for conditional generative tasks, effectively embedding condition constraints in the learning process.

Experimental Validation

The IGD framework was evaluated across three key domains:

  1. Layout Generation: IGD shows superior results in generating coherent layouts satisfying complex constraints, with evaluation metrics outperforming leading baseline models on datasets like PubLayNet and RICO. Key metrics included Frechet Inception Distance (FID) and Maximum Intersection over Union (mIoU).
  2. Molecule Generation: In handling the 3D structures in the QM9 dataset, IGD demonstrated impressive results in synthesizing molecules without relying on specialized equivariant architectures, maintaining high atom stability and molecular validity.
  3. Boolean Satisfiability (3-SAT): IGD achieved benchmark improvements in solving random 3-SAT instances, particularly noting scalable performance as the complexity of variable assignments increased. Figure 1

    Figure 1: Interleaved Noising Process: Sequential noising of discrete tokens (D_s) and continuous vectors (C_s). Noising occurs one element at a time, keeping other elements unchanged.

Conclusion

The Interleaved Gibbs Diffusion framework represents a significant advancement in generative modeling for discrete-continuous domains subject to implicit constraints. By sidestepping the need for denoising factorizability, IGD offers a robust approach to tackling complex, constrained generation tasks with empirical validations affirming its efficacy across diverse applications. Future work may explore optimizing sampler choices and integrating more specialized architectures and loss functions tailored to niche applications in layout and molecular generation.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.