Papers
Topics
Authors
Recent
Search
2000 character limit reached

π-Distill: QCD & PDE-Constrained Diffusion

Updated 6 February 2026
  • π-Distill is a dual framework that applies low-rank projection for high-precision lattice QCD and physics-informed diffusion model distillation.
  • In lattice QCD, it employs Laplacian eigenmode projection to create distilled quark fields, enhancing multi-hadron spectroscopy and energy extraction via variational analysis.
  • In physics-inspired machine learning, the PIDDM approach distills a teacher diffusion model into a single-step generator by enforcing PDE residuals for accurate system modeling.

π-Distill refers to two distinct methodologies developed for (1) high-precision correlation measurements in lattice quantum chromodynamics (QCD), and (2) the efficient, physically-constrained distillation of score-based diffusion models for generative modeling of systems governed by partial differential equations (PDEs). Both domains exploit a common principle: low-rank or post-hoc projection in order to enforce structural or physical constraints, either on quantum field correlators or on neural generative models.

1. Distillation for Lattice QCD: The π-Distill/Distillation Framework

The distillation method in lattice QCD constructs low-rank projectors onto the subspace of low-lying eigenmodes of the gauge-covariant Laplacian on each time-slice. These projectors are used to define smeared (“distilled”) quark fields, yielding all-to-all quark propagators (or “perambulators”) with significantly improved operator overlap and statistical quality for multi-hadron spectroscopy.

Distillation Operator Construction

On each time-slice tt, the three-dimensional gauge-covariant Laplacian xy2(t)-\nabla^2_{xy}(t) is diagonalized: xy2(t)ξy(k)(t)=λk(t)ξx(k)(t)-\nabla^2_{xy}(t)\,\xi^{(k)}_y(t) = \lambda_k(t)\,\xi^{(k)}_x(t) Ordering the eigenvalues λ1λ2\lambda_1 \leq \lambda_2 \leq \cdots, the lowest NN eigenmodes define the distillation space: V(t)=[ξ(1)(t),,ξ(N)(t)]V(t) = [\xi^{(1)}(t), \dots, \xi^{(N)}(t)] The rank-NN distillation projector is

P(t)=V(t)V(t)P(t) = V(t) V(t)^\dagger

The smeared (distilled) quark field is then given by

ψs(x,t)=[P(t)ψ(,t)]x\psi_s(x,t) = [P(t)\psi(\cdot, t)]_x

All creation and annihilation operators for composite hadrons are built from ψs\psi_s; the corresponding propagators form perambulators: τkl(t,0)=ξ(k)(t)D1(t,0)ξ(l)(0)\tau_{kl}(t,0) = \xi^{(k)\,\dagger}(t) D^{-1}(t,0) \xi^{(l)}(0) where DD is the Dirac operator (Woss et al., 2016, Lachini et al., 2021).

2. Extraction of Energy Spectra and Scattering Information

π-Distill provides the foundation for robust variational analysis using large operator bases. The correlator matrix

Cij(t)=Oi(t)Oj(0)C_{ij}(t) = \langle O_i(t) O_j^\dagger(0) \rangle

is constructed using interpolating operators projected onto definite lattice momenta and irreducible representations. A generalised eigenvalue problem (GEVP) is solved: C(t)vn(t)=λn(t,t0)C(t0)vn(t)C(t) v_n(t) = \lambda_n(t, t_0) C(t_0) v_n(t) with principal correlators λn(t,t0)\lambda_n(t, t_0) yielding the energies through multi-exponential fits.

Optimised “single-pion” and two-meson operators are constructed using GEVPs within each irrep, maximising overlap with target states. Combining these with Lüscher’s finite-volume method enables extraction of scattering phase shifts, particularly in channels dominated by elastic two-pion (ππ\pi\pi) states, e.g., for the extraction of the ρ resonance properties (Woss et al., 2016).

3. Implementation Choices: Rank Dependence, Stochastic and Exact Distillation

The choice of distillation rank (NN or NevN_{\rm ev}) controls both smearing radius and computational cost. Typically, 32N38432 \leq N \leq 384 (e.g., N=128N=128 on 32332^3 lattices) or Nev=64N_{\rm ev}=64 on a 483×9648^3\times96 lattice is used. The smearing profile

Ψ(r)=x,ttr[Sx,x+r(t)Sx+r,x(t)]\Psi(r) = \sum_{x,t} \sqrt{ \operatorname{tr}[ S_{x, x+r}(t) S_{x+r, x}(t) ] }

shows rapid saturation for Nev60N_{\rm ev} \gtrsim 60, with diminishing returns at higher ranks.

Two main approaches exist:

  • Exact distillation: Formation of all perambulators in the low-mode subspace, providing optimal error at moderate inversion cost (4Nev4N_{\rm ev} per source).
  • Stochastic distillation: Uses noise in the LapH subspace with dilution schemes, reducing the number of inversions but introducing stochastic noise, which requires normalization by cost to compare accuracy.

Empirically, exact distillation at Nev=64N_{\rm ev}=64 yields optimal performance for pion and KπK\pi correlation functions at physical masses, with consistent energies and efficient momentum projection. Stochastic distillation is suboptimal for this cost regime (Lachini et al., 2021).

Key implementation and performance characteristics established in the literature are summarized below.

Lattice Volume / Parameters π-Distill Key Choices Outcome / Recommendation
323×25632^3\times256, as0.12a_s\approx 0.12\,fm, mπ=236m_\pi=236 MeV N=128N=128 (4–10% spatial dim.), 20–35 qˉq\bar qq+5–6 ππ\pi\pi ops/irrep Energies/phase shifts converge at few-% level for N128N\geq128 (Woss et al., 2016)
483×9648^3\times96, a0.114a\simeq0.114 fm, mπ=139m_\pi=139 MeV Nev=64N_{\rm ev}=64, exact distillation Best cost–error trade-off; effective for large bases (Lachini et al., 2021)

Optimal practice is to select NN (NevN_{\rm ev}) such that the smearing radius covers the physical pion's size but avoids excessive computation. For the ππ\pi\pi channel up to the ρ resonance, convergence is achieved for ranks 128\gtrsim128 on moderate volumes; at the physical point, Nev=64N_{\rm ev}=64 suffices.

5. Physics-Informed Distillation of Diffusion Models (PIDDM)

In generative modeling of physics-governed systems, π-Distill (“PIDDM”) refers to the post-hoc distillation of pretrained score-based diffusion models, with explicit enforcement of PDE residual constraints at the level of the one-step student model output. The method builds upon a teacher diffusion model (trained via score-matching on noisy data) and distills its iterative sampler into a single neural network, training with a combined regression-to-teacher and physics residual penalty: Ldistill(θ)=Lsample+λtrainLPDE\mathcal L_{\rm distill}(\theta') = \mathcal L_{\rm sample} + \lambda_{\rm train} \mathcal L_{\rm PDE} where

LPDE=EϵR(dθ(ϵ))2,R(x)=[F[u],B[a]]\mathcal L_{\rm PDE} = \mathbb E_{\epsilon} \| \mathcal R(d_{\theta'}(\epsilon)) \|^2, \quad \mathcal R(x) = [\mathcal F[u], \mathcal B[a]]^\top

with physical constraints F,B\mathcal F, \mathcal B specified for the PDE class (Zhang et al., 28 May 2025).

Unlike earlier approaches, which enforced constraints on the score-model’s posterior mean (subject to Jensen's gap), PIDDM eliminates this gap by applying penalties solely to actual (fully generated) clean samples. This supports strict PDE satisfaction for forward, inverse, and partial observation generation.

6. Algorithmic Procedure, Experimental Validation, and Limitations

PIDDM operates in two stages:

  • Distillation: Generate a (noise, teacher sample) dataset, train the student to regress to teacher output plus minimize PDE residual.
  • Inference: For constrained tasks, refinement iterations may further penalize PDE violation, though single-step generation already achieves competitive accuracy.

Validation over benchmarks (Darcy flow, Poisson, Burgers’, etc.) demonstrates that PIDDM achieves significantly reduced PDE error (0.150.2×104\sim 0.15-0.2\times10^{-4}, NFE as low as 1) compared to diffusion models with stepwise enforcement. Ablations indicate that higher teacher fidelity and moderate loss tradeoffs optimize sample quality and physics accuracy. The method is limited by reliance on a well-trained teacher and a discretized PDE operator. Future directions include embedding constraints at the architectural/tokenizer level, handling stiffer equations, and scaling to 3D/multiphysics scenarios (Zhang et al., 28 May 2025).

7. Significance and Broader Context

Distillation-based frameworks, as exemplified by π-Distill, enable efficient, low-noise extraction of physically meaningful observables—whether in precision lattice spectroscopy (via Laplacian eigen-projection and variational analysis) or fast, physical constraint-satisfying generative modeling (via post-hoc one-step network distillation). For lattice QCD, the method is now a standard tool for large-volume, high-precision studies of multi-hadron systems; for physics-inspired machine learning, PIDDM provides a principled solution to enforcing hard physical constraints in generative pipelines. Both applications illustrate the general efficacy of low-rank or sample-level projection for balancing fidelity, tractability, and structure in data-driven modeling.

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to π-Distill.