Papers
Topics
Authors
Recent
Search
2000 character limit reached

Learned Multi-Layer VAMP (LMLVAMP)

Updated 2 February 2026
  • LMLVAMP is a hybrid estimator that combines model-based inference with data-driven neural denoisers to recover signals from complex, nonlinear, and quantized measurements.
  • It integrates spectral priors with neural network–based denoising in a message-passing framework, leveraging fast Fourier transforms for efficient processing.
  • The approach achieves significant NMSE reductions and approaches oracle performance in FR3 6G scenarios, effectively mitigating interference, saturation, and quantization effects.

Learned Multi-Layer Vector Approximate Message Passing (LMLVAMP) is a hybrid model-based/data-driven estimator designed for recovering a desired signal from nonlinear, quantized receiver observations in the presence of strong out-of-band (OOB) interference, front-end saturation nonlinearities, additive receiver noise, and finite-resolution quantization. LMLVAMP integrates spectral priors with neural network–based denoising within a principled message-passing algorithmic structure. Applications include upper mid-band (FR3, 7–24 GHz) wideband radio receivers for 6G, where spectrally separated interference and hardware-induced nonlinear distortions degrade performance beyond the capacity of conventional linear methods (Joy et al., 30 Jan 2026).

1. System and Observation Model

The system considers NN-length time-domain samples rCNr \in \mathbb{C}^N comprising a superposition of LL frequency-domain sources:

r==0L1Vx,r = \sum_{\ell=0}^{L-1} V x_\ell,

where VV is the N×NN \times N unitary inverse discrete Fourier transform (IDFT), and each x[k]x_\ell[k] is nonzero only within its designated frequency band BB_\ell, with BBk=B_\ell \cap B_k = \emptyset for k\ell \ne k.

The frequency-domain coefficients xx_\ell are assigned Gaussian priors:

x[i]{N(μ[i],S)iB, 0otherwise.x_\ell[i] \sim \begin{cases} \mathcal{N}(\mu_\ell[i], S_\ell) & i \in B_\ell, \ 0 & \text{otherwise.} \end{cases}

In typical use, =0\ell=0 denotes the desired user, and =1\ell=1 indicates a spectrally separated interferer.

The receiver front-end applies a smooth, memoryless amplitude-compressing nonlinearity, followed by additive white Gaussian noise and, optionally, scalar quantization:

y[i]=ϕ(r[i],wa[i],wb[i])=f(r[i]+wa[i]Psat)(r[i]+wa[i])+wb[i],y[i] = \phi(r[i], w_a[i], w_b[i]) = f\left( \frac{|r[i] + w_a[i]|}{\sqrt{P_{\rm sat}}} \right)(r[i] + w_a[i]) + w_b[i],

with

f(x)=tanh(x)x,f(x) = \frac{\tanh(x)}{x},

wa[i]CN(0,σa2)w_a[i] \sim \mathcal{CN}(0, \sigma_a^2) (pre-nonlinearity noise), wb[i]CN(0,σb2)w_b[i] \sim \mathcal{CN}(0, \sigma_b^2) (post-nonlinearity noise), and PsatP_{\rm sat} the saturation threshold. For finite-resolution analog-to-digital conversion (ADC), a uniform scalar quantizer Q\mathcal{Q} is applied: yq[i]=Q(y[i])y_q[i] = \mathcal{Q}(y[i]).

2. Multi-Layer VAMP Algorithmic Structure

Signal recovery is formulated as inference in a two-layer network:

  • Layer 0 (Spectral): xr=Vxx \mapsto r = Vx
  • Layer 1 (Nonlinear): ry=Φ(r,w)r \mapsto y = \Phi(r, w)

The classical Multi-Layer Vector Approximate Message Passing (ML-VAMP) algorithm alternates two denoising steps per iteration, transitioning between frequency and time domains via orthogonal transforms. The denoisers are augmented by Onsager-like corrections for improved convergence. The Bayesian ML-VAMP update sequence is: x(t)=G0(z0(t),γ0(t)),α0(t)=G0, z1(t)=V(x(t)α0(t)z0(t))1α0(t),γ1(t)=γ0(t)(1/α0(t)1), r(t)=G1(z1(t),γ1(t)),α1(t)=G1, z0(t+1)=V(r(t)α1(t)z1(t))1α1(t),γ0(t+1)=γ1(t)(1/α1(t)1).\begin{aligned} x^{(t)} &= G_0(z_0^{(t)}, \gamma_0^{(t)}), \quad \alpha_0^{(t)} = \langle G_0' \rangle, \ z_1^{(t)} &= \frac{V(x^{(t)} - \alpha_0^{(t)} z_0^{(t)})}{1 - \alpha_0^{(t)}}, \quad \gamma_1^{(t)} = \gamma_0^{(t)}(1/\alpha_0^{(t)} - 1), \ r^{(t)} &= G_1(z_1^{(t)}, \gamma_1^{(t)}), \quad \alpha_1^{(t)} = \langle G_1' \rangle, \ z_0^{(t+1)} &= \frac{V(r^{(t)} - \alpha_1^{(t)} z_1^{(t)})}{1 - \alpha_1^{(t)}}, \quad \gamma_0^{(t+1)} = \gamma_1^{(t)}(1/\alpha_1^{(t)} - 1). \end{aligned} Here, G0G_0 is the spectral (linear-Gaussian) denoiser, yielding, for each frequency bin ii:

x[i]={μ[i]+γ0S1+γ0S(z0[i]μ[i])iB, 0iB,x[i] = \begin{cases} \mu_\ell[i] + \frac{\gamma_0 S_\ell}{1+\gamma_0 S_\ell}(z_0[i] - \mu_\ell[i]) & i \in B_\ell, \ 0 & i \notin \cup_\ell B_\ell, \end{cases}

with average divergence α0=BNγ0S1+γ0S\alpha_0 = \sum_\ell \frac{|B_\ell|}{N} \frac{\gamma_0 S_\ell}{1+\gamma_0 S_\ell}. The nonlinear denoiser G1G_1 is the componentwise conditional mean estimator E{ry=ϕ(r,w),rCN(z1,γ11)}\mathbb{E}\{ r \mid y = \phi(r, w), r \sim \mathcal{CN}(z_1, \gamma_1^{-1}) \}, which lacks a closed-form in presence of saturation and quantization.

3. Learned Neural Network Denoisers

LMLVAMP generalizes ML-VAMP by replacing analytic denoisers G0G_0 and G1G_1 with small, trainable neural network denoisers f0f_0 and f1f_1. These networks learn to emulate minimum mean-squared error (MMSE) properties and estimate Onsager-correcting divergences.

  • Spectral Message Updater (f0f_0): For each frequency bin ii at iteration tt,

ρ0(t)[i]=f0(z0(t)[i],γ0(t),S[i],μ[i],t;θ)(β~0[i],β~1[i]),\rho_0^{(t)}[i] = f_0(z_0^{(t)}[i], \gamma_0^{(t)}, S[i], \mu[i], t;\theta) \to (\tilde\beta_0[i], \tilde\beta_1[i]),

with input features [z0/Psat,(γ0Psat)1,S/Psat,μ/Psat][|z_0|/\sqrt{P_{\rm sat}}, (\gamma_0 P_{\rm sat})^{-1}, S/P_{\rm sat}, |\mu|/\sqrt{P_{\rm sat}}]. The iteration-wide coefficients are

(β0(t),β1(t))=1Ni=0N1ρ0(t)[i].(\beta_0^{(t)}, \beta_1^{(t)}) = \frac{1}{N} \sum_{i=0}^{N-1} \rho_0^{(t)}[i].

  • Nonlinear Denoiser (f1f_1): For each sample ii,

(v(t)[i],ρ1(t)[i])=f1(z1(t)[i],γ1(t),y[i],t;θ),(v^{(t)}[i], \rho_1^{(t)}[i]) = f_1(z_1^{(t)}[i], \gamma_1^{(t)}, y[i], t;\theta),

with features [z1/Psat,(γ1Psat)1,y/Psat][|z_1|/\sqrt{P_{\rm sat}}, (\gamma_1 P_{\rm sat})^{-1}, |y|/\sqrt{P_{\rm sat}}]. Both f0f_0 and f1f_1 are two-layer networks (64 sigmoid units with linear outputs).

4. Algorithmic Workflow and Pseudocode

The LMLVAMP inference procedure unrolls for TT iterations as follows:

  1. Initialization: z1(0)Vμz_1^{(0)} \leftarrow V \mu, γ1(0)S1\gamma_1^{(0)} \leftarrow S^{-1}.
  2. For t=0,...,T1t=0,...,T-1:
    • Nonlinear denoising: For all ii, (v[i],ρ1[i])f1(z1[i],γ1,y[i],t;θ)(v[i], \rho_1[i]) \leftarrow f_1(z_1[i], \gamma_1, y[i], t; \theta).
    • Spectral transformation: z0Vvz_0 \leftarrow V v, γ0(1/N)iρ1[i]\gamma_0 \leftarrow (1/N) \sum_i \rho_1[i].
    • Spectral denoising: xμ+(γ0S1+γ0S)(z0μ)x \leftarrow \mu + \left(\frac{\gamma_0 S}{1+\gamma_0 S}\right) \odot (z_0 - \mu).
    • Update: γ1γ0(1Niγ0S[i]1+γ0S[i])1\gamma_1 \leftarrow \gamma_0 \cdot \left(\frac{1}{N} \sum_i \frac{\gamma_0 S[i]}{1+\gamma_0 S[i]}\right)^{-1}.
    • Message update: For all ii, ρ0[i]f0(z0[i],γ0,S[i],μ[i],t;θ)\rho_0[i] \leftarrow f_0(z_0[i], \gamma_0, S[i], \mu[i], t;\theta).
    • Aggregate: (β0,β1)(1/N)iρ0[i](\beta_0, \beta_1) \leftarrow (1/N) \sum_i \rho_0[i].
    • Next iterate: z1β0Vxβ1Vz0z_1 \leftarrow \beta_0 V x - \beta_1 V z_0.
  3. User-band selection: x^0mB0x\hat{x}_0 \leftarrow m_{B_0} \odot x.

Forward and inverse FFT operations are leveraged for computational efficiency (O(N)O(N) per iteration).

5. Training Objectives and Optimization

Trainable parameters θ\theta of f0f_0 and f1f_1 are optimized end-to-end via backpropagation through the TT-step unrolled LMLVAMP. The total loss function is a convex combination of:

  • Early-iteration loss:

Learly=t=1T1wtx0[B0]x^(t)[B0]2,wt=ti=1T1i\mathcal{L}_{\rm early} = \sum_{t=1}^{T-1} w_t \| x_0[B_0] - \hat{x}^{(t)}[B_0] \|^2, \qquad w_t = \frac{t}{\sum_{i=1}^{T-1} i}

  • Final-iteration loss: Lfinal=x0[B0]x^(T)[B0]2\mathcal{L}_{\rm final} = \| x_0[B_0] - \hat{x}^{(T)}[B_0] \|^2
  • Total loss: Ltotal=ηLfinal+(1η)Learly\mathcal{L}_{\rm total} = \eta \mathcal{L}_{\rm final} + (1-\eta) \mathcal{L}_{\rm early} with η(0.5,1]\eta \in (0.5, 1]

Optimization employs Adam with exponential learning-rate decay. Regularization is enforced by restricting network size and introducing weighted intermediate losses for training stability.

6. Performance and Evaluation in FR3 Coexistence Scenarios

Simulated configurations included N=512N=512, B0=[0,100)B_0 = [0, 100), B1=[300,400)B_1 = [300, 400), pre-nonlinearity noise σa2=0\sigma_a^2 = 0 dB, post-nonlinearity noise σb2=10\sigma_b^2 = -10 dB, saturation SNR =40=40 dB, signal SNR {10,20}\in \{10, 20\} dB, interference-to-noise ratio (INR) [30,80]\in [30, 80] dB, and 10-bit quantization with 12 dB backoff. Competing estimators were LMLVAMP-K/U (with/without known interferer band), linear Wiener baselines, and an "oracle" ideal nonlinearity-inversion bound.

Metrics:

  • Achievable rate lower bound: Clog2(1ρ)C \geq -\log_2(1-\rho) (correlation coefficient ρ\rho)
  • Normalized MSE

Key results:

  • LMLVAMP-K approached oracle rates within two iterations at INR =80=80 dB.
  • LMLVAMP-U achieved >>20 dB NMSE reduction compared to linear methods in saturation-dominated scenarios.
  • With 10-bit quantization, LMLVAMP retained a >>10 dB advantage over linear approaches.
  • Additional algorithmic unfolding (iterations) yielded incremental performance improvements.
  • Ablations fixing (β0,β1)=(1,0)(\beta_0, \beta_1) = (1, 0) had negligible performance loss, indicating stable convergence.

LMLVAMP's O(NN) per-iteration complexity and compact parameterization (small neural nets) demonstrate scalability to large systems and robust gains in realistic 6G coexistence conditions (Joy et al., 30 Jan 2026).

7. Significance and Implications

LMLVAMP exemplifies the hybridization of model-based inference (structured priors, orthogonal transforms) and machine-learned components (data-driven denoisers) within message-passing frameworks. By incorporating spectral priors and leveraging neural denoisers to bypass intractable conditional expectations due to nonlinearities and quantization, LMLVAMP successfully bridges practical front-end hardware constraints and advanced signal recovery objectives. This architecture is particularly suitable for wideband communications in spectrally dense environments, such as future FR3 6G scenarios, where classical linear estimators fail to address nonlinear spectral leakage and quantization artifacts.

A plausible implication is the extensibility of LMLVAMP principles to other nonlinear or quantized inference tasks in communications and signal processing, given its stability, computational efficiency, and end-to-end trainability.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (1)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Learned Multi-Layer Vector Approximate Message Passing (LMLVAMP).