Papers
Topics
Authors
Recent
Search
2000 character limit reached

Coordinate-Conditioned Denoising Diffusion Model

Updated 3 January 2026
  • CCDDPM is a generative framework that produces high-fidelity, scene-consistent radio environment maps conditioned on transmitter coordinates.
  • It fuses a coordinate-derived Gaussian prior with Gaussian noise in a two-channel U-Net to accurately emulate dynamic 6G vehicular radio conditions.
  • The model outperforms traditional methods by delivering real-time REM synthesis with low uncertainty and preserved spatial details essential for C-V2X communications.

A Coordinate-Conditioned Denoising Diffusion Probabilistic Model (CCDDPM) is a generative framework designed to synthesize high-fidelity, scene-consistent radio environment maps (REMs) for 6G Cellular Vehicle-to-Everything (C-V2X) communications as a function of arbitrary transmitter vehicle coordinates. In this architecture, conditional generative diffusion processes are leveraged to predict dense two-dimensional received-power fields, enabling rapid and accurate emulation of radio propagation environments under dynamic vehicular transmitter locations, without recourse to exhaustive measurement campaigns or computationally expensive ray tracing. CCDDPM achieves this by fusing a smooth, coordinate-derived Gaussian prior with Gaussian noise inside a lightweight, two-channel conditional U-Net, which is trained end-to-end to reverse the noise process and recover credible REM samples conditioned on spatial context (Cao et al., 27 Dec 2025).

1. Motivation and Background

The 6G C-V2X paradigm necessitates dynamic, fine-grained REMs, denoted x0RH×Wx_0 \in \mathbb{R}^{H \times W}, to support real-time communication reliability, handover, and scheduling as vehicles traverse urban environments. Traditional REM acquisition is labor-intensive, relying on active drive tests or computationally intensive ray-tracing models. The spatially non-IID nature of radio propagation—driven by continuous vehicular movement and complex urban architectures—renders classical generative strategies (e.g., VAE, GAN, NF) insufficient for capturing both global patterns and transmitter-centric local effects. CCDDPM addresses this challenge by learning the conditional distribution p(x0c)p(x_0 \mid c), where c=(x0,y0)c = (x_0, y_0) denotes the transmitter coordinate, enabling synthesis of REMs for any queried transmitter location in a given region. This approach provides statistical and structural fidelity while supporting real-time deployment on edge hardware (Cao et al., 27 Dec 2025).

2. Mathematical Foundations

The CCDDPM framework generalizes the denoising diffusion probabilistic model to a coordinate-conditioned setting:

  • Forward (Noising) Process

    • Initiated from a clean REM x0x_0, Gaussian noise is added in TT steps:
    • q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T
    • where βt{\beta_t} is a predetermined noise schedule; αt1βt\alpha_t \equiv 1 - \beta_t; cumulative αˉts=1tαs\bar\alpha_t \equiv \prod_{s=1}^t \alpha_s.
    • The process admits a closed-form sampling at any step:

    q(xtx0)=N(xt;αˉtx0,  (1αˉt)I)q(x_t \mid x_0) = \mathcal{N}\left(x_t; \sqrt{\bar\alpha_t} x_0,\; (1-\bar\alpha_t) I \right)

  • Reverse (Denoising) Process

    • A parameterized conditional model p(x0c)p(x_0 \mid c)0 in Gaussian form:

    p(x0c)p(x_0 \mid c)1 - Mean prediction is cast in terms of the injected noise p(x0c)p(x_0 \mid c)2:

    p(x0c)p(x_0 \mid c)3

  • Optimization Objective

    • The evidence lower bound (ELBO) on p(x0c)p(x_0 \mid c)4 is decomposed into denoising sub-losses:

    p(x0c)p(x_0 \mid c)5 - The overall training loss averages over p(x0c)p(x_0 \mid c)6:

    p(x0c)p(x_0 \mid c)7

3. Coordinate Conditioning Mechanism

The transmitter coordinate, p(x0c)p(x_0 \mid c)8, is represented as a 2D Gaussian heatmap over the REM grid:

p(x0c)p(x_0 \mid c)9

with c=(x0,y0)c = (x_0, y_0)0 controlling Gaussian spread. This spatial prior is normalized to c=(x0,y0)c = (x_0, y_0)1, then concatenated with the noisy REM c=(x0,y0)c = (x_0, y_0)2, yielding a two-channel input c=(x0,y0)c = (x_0, y_0)3 for the denoising network. This explicit fusion anchors the generative process to the physical transmitter position throughout all denoising stages.

4. Model Architecture

The Denoiser adopts a lightweight two-channel U-Net architecture with temporal and conditional modulation:

  • Inputs: Two-channel tensor c=(x0,y0)c = (x_0, y_0)4, timestep embedding c=(x0,y0)c = (x_0, y_0)5
  • Time Embedding: Sinusoidal function c=(x0,y0)c = (x_0, y_0)6 mapped via an MLP to c=(x0,y0)c = (x_0, y_0)7; at each block, c=(x0,y0)c = (x_0, y_0)8 is used in FiLM (Feature-wise Linear Modulation) format:

c=(x0,y0)c = (x_0, y_0)9

  • Encoder: Repeats blocks of Conv–GroupNorm–SiLU, with FiLM-modulated residuals and x0x_00 downsampling; channel progression typically x0x_01; one multi-head self-attention block at x0x_02.
  • Bottleneck: Two residual blocks interleaved with self-attention.
  • Decoder: Upsampling by 2, concatenating encoder features via skip connections, followed by FiLM-modulated residuals, mirroring the encoder to reach the single-channel output.
  • Output: x0x_03 convolution to predict x0x_04.

5. Training and Inference Procedures

The CCDDPM training and sampling pipelines proceed as follows:

Stage Procedure Output
Training 1. Sample REM mini-batch x0x_05, coordinates x0x_06 x0x_07, x0x_08
2. Build, normalize x0x_09 TT0
3. Sample TT1, TT2 TT3, TT4
4. Form TT5 TT6
5. Predict TT7 TT8
6. Compute loss TT9, update q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T0 model update
Inference 1. Given q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T1, build q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T2 q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T3
2. Initialize q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T4 q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T5
3. For q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T6:
(a) q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T7
(b) q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T8 q(xtxt1)=N(xt;1βtxt1,βtI),t=1,...,Tq(x_t \mid x_{t-1}) = \mathcal{N}(x_t; \sqrt{1-\beta_t}\, x_{t-1}, \beta_t I), \quad t = 1, ..., T9
(c) βt{\beta_t}0; βt{\beta_t}1; βt{\beta_t}2 βt{\beta_t}3
4. Return βt{\beta_t}4, remapped to original signal-strength scale Synthesized REM

This process supports few-step samplers (e.g., DDIM, DPMSolver with βt{\beta_t}5) for efficient, real-time REM synthesis on edge GPUs (Cao et al., 27 Dec 2025).

6. Experimental Evaluation

Evaluation was conducted on 900 training and 100 test REMs (βt{\beta_t}6 m resolution):

  • Distributional Fidelity: At coordinate (108,178), the CDF of signal intensity from 100 generated REMs matches the empirical CDF closely, surpassing normalizing flows, GAN, and VAE.
  • Sampling Variance: Standard-deviation envelope over 100 samples is minimum for CCDDPM, indicating stability.
  • Line-slice RMSE: At βt{\beta_t}7 and βt{\beta_t}8, CCDDPM minimizes mean RMSE and error bars, especially at building edges and rapid spatial transitions.
  • Qualitative Structure: REMs synthesized via CCDDPM preserve hotspots, shadowing, and structural features (e.g., building outlines); alternatives over-smooth, introduce artifacts, or drift in contrast.
  • Runtime: Benefiting from lightweight U-Net and compatibility with fast samplers, CCDDPM enables real-time REM generation on edge GPUs.

7. Broader Implications and Impact

Conditional Gaussian priors on transmitter coordinates allow CCDDPM to accurately anchor localized high-power zones and model both global and local radio propagation structures. Key outcomes include:

  • High-fidelity REM Prediction: Preservation of spatial fine structure critical for radio planning and adaptive vehicular communication.
  • Consistent, Low-Uncertainty Samples: Repeatability and stability facilitate targeted active re-measurement in regions of highest epistemic uncertainty.
  • Operational Utility: On-the-fly REM inference supports enhanced PHY reliability, adaptive scheduling, and efficient handover for 6G C-V2X, eliminating the need for pervasive measurements.

A plausible implication is that this approach generalizes to other spatially conditioned generative tasks involving continuous context variables and complex scene structures. Coordinate conditioning thus converts generic diffusion into an efficient, scene-consistent REM predictor underpinning robust radio-map services for next-generation vehicular networks (Cao et al., 27 Dec 2025).

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Coordinate-Conditioned Denoising Diffusion Probabilistic Model (CCDDPM).