The paper introduces a GPR method that recovers full channel state information from partial, noise-corrupted pilots with MMSE optimality.
It leverages advanced covariance kernels—including spatial, data-adaptive, and geometry-aware mixtures—to capture channel correlations and reduce pilot overhead.
The framework provides calibrated uncertainty estimates and scalable computation, addressing practical challenges in large-scale multi-antenna systems.
A GPR-based channel estimation framework leverages Gaussian process regression to recover complete channel state information (CSI) in large-scale MIMO or multi-antenna wireless systems from partial or subsampled, noise-corrupted pilot observations. These frameworks model the spatially or space-time correlated fading channel as a realization of a complex-valued Gaussian process over antenna arrays, with kernels embedding the geometry and physical propagation structure. Posterior inference yields closed-form minimum mean-square error (MMSE) estimates and provides calibrated uncertainty quantification, enabling substantial reduction of pilot overhead and improved spectral efficiency compared to classical schemes.
1. Channel and Observation Models
The fundamental setting assumes a narrowband MIMO system with Nt transmit and Nr receive antennas. The instantaneous channel matrix is H∈CNr×Nt, vectorized as u=vec(H)∈CM, M=NrNt. Pilot resources are economized by exciting only a subset nt<Nt of transmit antennas, producing the observation model
y=Bu+ε,ε∼CN(0,σ2IP),P=Nrnt,
where B selects the sounded entries of u. The estimation goal is full recovery of u (hence H) from y. Key metrics include normalized mean-square error (NMSE), empirical 95% credible-interval coverage, and post-equalization spectral efficiency (SE) computed with the estimated channel (Shah et al., 21 Jan 2026, Shah et al., 27 Dec 2025, Shah et al., 29 Oct 2025).
2. Gaussian Process Regression Formulation
Each channel matrix coefficient Hr,t is modeled as the value of a latent complex-valued function f:G→C on a discrete antenna index set G, under a proper zero-mean GP prior
f(x)∼GP(0,k(x,x′)),x,x′∈G.
Observed entries {yi} arise via noisy sampling yi=f(xi)+εi at training points xi∈X⊂G; the remaining entries are inferred at X∗=G∖X. The GP prior is specified by a covariance function or kernel k, which encodes spatial correlation, array geometry, or statistical channel knowledge (Shah et al., 27 Dec 2025, Shah et al., 21 Jan 2026). The posterior distribution over the unobserved entries is analytically tractable, with mean and covariance as detailed below.
3. Covariance Kernel Design
Three principal paradigm classes arise in recent works:
Spatial-Correlation (SC) Kernel: Uses the known theoretical or empirical second-order statistics of the channel, with
kSC((r,t),(r′,t′))=[RH]n,m=E[Hr,tHr′,t′∗],
where RH is the full channel covariance, producing a kernel that faithfully reproduces transmit–receive coupling without auxiliary hyperparameters (Shah et al., 21 Jan 2026).
Data-Adaptive Kernels: Employ learned parameterized functions using array locations, such as
Radial basis function (RBF): kRBF(x,x′)=σf2exp(−∥x−x′∥2/2ℓ2),
Matérn: $k_{\mathrm{Mat with explicit smoothness/hyperparameters,</li>
<li>Rational quadratic (RQ): for multi-scale variability,</li>
<li>with hyperparameters learned from data by maximizing the marginal likelihood (<a href="/papers/2510.25390" title="" rel="nofollow" data-turbo="false" class="assistant-link" x-data x-tooltip.raw="">Shah et al., 29 Oct 2025</a>).</li>
</ul></li>
<li><strong>Geometry-Based Spectral Mixture (GB-SMCF):</strong> Constructs a separable kernel reflecting the spatial structure of physical antenna placements</li>
</ul>
<p>$k_\mathrm{base}((i,j),(i',j');\theta)
= A\,k_r(i,i')\,k_t(j,j')</p><p>witheachk_sasumofcomplex2Dspectralmixturecomponentsmodelingclusteredangularstatistics.Physicalantennacoordinatesareexplicitlyencoded,andallkernelandcoregionalizationhyperparametersarejointlyoptimizedonline(<ahref="/papers/2512.22578"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,27Dec2025</a>).</p><h2class=′paper−heading′id=′posterior−inference−and−mmse−optimality′>4.PosteriorInferenceandMMSEOptimality</h2><p>GivenPnoisyobservationsindexedby\mathbf{X}_OandMpredictionlocations\mathbf{X}_*,<ahref="https://www.emergentmind.com/topics/generalized−pseudo−label−robust−gpr−loss"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">GPR</a>yields</p><p>\hat{\mathbf{h}}_* = \mu_{\rm post} = K_{*O} (K_{OO}+\sigma^2 I_P)^{-1} \mathbf{y}</p><p>\Sigma_{\rm post} = K_{**} - K_{*O} (K_{OO}+\sigma^2 I_P)^{-1} K_{O*}</p><p>withK_{OO},K_{*O},K_{**}constructedfromthekernelevaluatedonobservedandtestpoints.</p><p>FortheSCkernel,theGPRposteriormeanexactlycoincideswiththeclassicallinearMMSEestimatorunderthegivensecond−orderstatistics,establishingMMSEoptimalityregardlessofunderlyingchannelGaussianity(<ahref="/papers/2601.14759"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,21Jan2026</a>,<ahref="/papers/2510.25390"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,29Oct2025</a>).Whenthekernelislearnedfromdata,theposteriormeancorrespondstothebestlinearunbiasedpredictor(BLUP)forgeneral,potentiallynon−Gaussian,second−ordermodels(<ahref="/papers/2510.25390"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,29Oct2025</a>,<ahref="/papers/2512.22578"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,27Dec2025</a>).</p><h2class=′paper−heading′id=′pilot−reduction−complexity−and−uncertainty−quantification′>5.PilotReduction,Complexity,andUncertaintyQuantification</h2><p>GPR−basedschemespermitaggressivepilotoverheadreductionwhilemaintainingaccuracyandcomputationaltractability.ThedominantcomputationalcostistheinversionofaP\times Pmatrix,scalingas\mathcal{O}(P^3)withP=N_{\mathrm{r}}n_{\mathrm{t}}\ll MN,substantiallylowerthanfull−dimensionalMMSEschemes(<ahref="/papers/2601.14759"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,21Jan2026</a>,<ahref="/papers/2512.22578"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,27Dec2025</a>,<ahref="/papers/2510.25390"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,29Oct2025</a>).Table1summarizesempiricalresultsfortypicalbenchmarkscenarios(<ahref="/papers/2601.14759"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,21Jan2026</a>).</p><divclass=′overflow−x−automax−w−fullmy−4′><tableclass=′tableborder−collapsew−full′style=′table−layout:fixed′><thead><tr><th>Estimator</th><thstyle="text−align:right">Pilotsavings</th><thstyle="text−align:right">NMSE[dB]</th><thstyle="text−align:right">RelativeSE[<thstyle="text−align:left">Complexity</th></tr></thead><tbody><tr><td>SC−GPR(\Delta=2)</td><tdstyle="text−align:right">50<tdstyle="text−align:right">–14.75</td><tdstyle="text−align:right">94.5</td><tdstyle="text−align:left">\mathcal{O}(648^3)</td></tr><tr><td>RBF−GPR(\Delta=2)</td><tdstyle="text−align:right">50<tdstyle="text−align:right">–2.81</td><tdstyle="text−align:right">76.1</td><tdstyle="text−align:left">\mathcal{O}(Q\cdot648^3)</td></tr><tr><td>MMSE(full)</td><tdstyle="text−align:right">0<tdstyle="text−align:right">–10.49</td><tdstyle="text−align:right">73.9</td><tdstyle="text−align:left">\mathcal{O}(1296^3)</td></tr></tbody></table></div><p>Empirical95<h2class=′paper−heading′id=′kernel−choices−hyperparameter−optimization−and−practical−guidelines′>6.KernelChoices,HyperparameterOptimization,andPracticalGuidelines</h2><p>Kernelselectioncriticallyaffectsperformance,especiallyforanisotropicorundersampledantennaconfigurations.Inregular2Darrayscenarios,Euclideandistance−basedkernels(RBF,Mateˊrn,RQ)areeffective,butwithsparse,directional,ordiagonalsampling,MateˊrnandRQkernels(allowingrougherstructure)outperformRBF.Geometry−aware<ahref="https://www.emergentmind.com/topics/spectral−mixture−sm−kernels"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">spectralmixturekernels</a>provideinterpretable,physicallygroundedparameterizationsandenableenergy−efficientadaptivelearningwithonlinehyperparametertuning(<ahref="/papers/2512.22578"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,27Dec2025</a>).</p><p>Alldata−drivenkernelsemploygradient−basedoptimization(e.g.,L−BFGS)ofthelog−marginallikelihood,withcomputationandmemorycomplexitydominatedbytheCholeskyfactorizationofK_{OO}+\sigma^2 I.</p><p>Forscalabilityandlarge−scalearrays,onemayexploitKroneckerorToeplitzstructure,inducing−pointsparseapproximations,orconjugate−gradientbasedsolvers,leveragingmatrix−vectorproductacceleration(<ahref="/papers/2510.25390"title=""rel="nofollow"data−turbo="false"class="assistant−link"x−datax−tooltip.raw="">Shahetal.,29Oct2025</a>).</p><h2class=′paper−heading′id=′performance−analysis−and−extensions′>7.PerformanceAnalysisandExtensions</h2><p>SimulationsusingrealisticmmWavearraydimensions(e.g.,36\times36$) and channel models (Kronecker, Weichselberger, Saleh–Valenzuela, geometry-based clustered) confirm:
Pilot reduction up to 75% is attainable, incurring only moderate NMSE and SE degradation, exceeding performance of LS/MMSE at moderate and low SNRs.
GPR schemes systematically produce well-calibrated posterior uncertainties and are robust to non-Gaussian channel statistics.
Proposed frameworks are readily extensible to multi-user MIMO (via multi-output GPs), spatio-temporal online tracking, wideband/frequency-selective channels (by augmenting the kernel domain), and hybrid analog-digital hardware constraints by altering the observation operator (Shah et al., 27 Dec 2025).
8. Assumptions, Limitations, and Future Directions
Present methods rely on either knowledge of the second-order covariance matrix or the capacity to learn spatial kernel hyperparameters from limited data. For the SC kernel approach, knowledge or consistent estimation of RH is assumed. Data-driven methods mitigate this by nonparametric kernel learning, but at cubic training cost per block, which is partially offset by structural exploitation and sparse approximations for very large P.
Hyperparameter identifiability is improved through concise box constraints, initialization, and, in practice, regularization. The Gaussian process foundational assumption enables robust uncertainty quantification and principled interpolation but may require adaptation for non-stationary or highly dynamic propagation environments.
A plausible implication is that, as array sizes and operable bandwidths increase, GPR-based frameworks—particularly those embedding explicit physical/geometry-aware priors—will form an essential component of efficient, reliable, and energy-effective multi-antenna channel estimation systems (Shah et al., 21 Jan 2026, Shah et al., 27 Dec 2025, Shah et al., 29 Oct 2025).
Pilot reduction up to 75% is attainable, incurring only moderate NMSE and SE degradation, exceeding performance of LS/MMSE at moderate and low SNRs.
GPR schemes systematically produce well-calibrated posterior uncertainties and are robust to non-Gaussian channel statistics.
Proposed frameworks are readily extensible to multi-user MIMO (via multi-output GPs), spatio-temporal online tracking, wideband/frequency-selective channels (by augmenting the kernel domain), and hybrid analog-digital hardware constraints by altering the observation operator (Shah et al., 27 Dec 2025).
8. Assumptions, Limitations, and Future Directions
Present methods rely on either knowledge of the second-order covariance matrix or the capacity to learn spatial kernel hyperparameters from limited data. For the SC kernel approach, knowledge or consistent estimation of RH is assumed. Data-driven methods mitigate this by nonparametric kernel learning, but at cubic training cost per block, which is partially offset by structural exploitation and sparse approximations for very large P.
Hyperparameter identifiability is improved through concise box constraints, initialization, and, in practice, regularization. The Gaussian process foundational assumption enables robust uncertainty quantification and principled interpolation but may require adaptation for non-stationary or highly dynamic propagation environments.
A plausible implication is that, as array sizes and operable bandwidths increase, GPR-based frameworks—particularly those embedding explicit physical/geometry-aware priors—will form an essential component of efficient, reliable, and energy-effective multi-antenna channel estimation systems (Shah et al., 21 Jan 2026, Shah et al., 27 Dec 2025, Shah et al., 29 Oct 2025).