Gaussian Kernel Feature Space
- Feature Space for Gaussian Kernels is the infinite-dimensional reproducing kernel Hilbert space induced by the Gaussian function, enabling inner product representations and universal approximations.
- Various constructions including Mercer, Hermite, Fourier, and Segal–Bargmann expansions provide distinct analytical frameworks based on eigen-decompositions and spectral methods.
- Practical techniques like random Fourier features and explicit finite-dimensional mappings offer efficient approximations and implementations in kernel-based statistical learning.
A Gaussian kernel is a symmetric, positive-definite function of the form
or, more generally, defined on a real separable Hilbert space . The feature space corresponding to a Gaussian kernel is the Hilbert space into which data are mapped such that equals the inner product between the mapped points. This feature space is typically infinite-dimensional, and its structure underlies the expressive power, universality, and practical implementation of kernel methods in statistical learning, Gaussian processes, SVMs, and related fields. Several analytically distinct but isometrically equivalent constructions capture the geometry and analytic regularity of the Gaussian kernel's feature space across finite- and infinite-dimensional domains.
1. Reproducing Kernel Hilbert Space of the Gaussian Kernel
Let be a real separable Hilbert space. The Gaussian kernel is positive definite on . By Aronszajn’s theorem, there exists a unique reproducing kernel Hilbert space (RKHS) consisting of functions with the following properties (Guella, 2020):
- For each ,
- The reproducing property: ,
- The span is dense in
Every admits a representation (converging in ), with norm . The feature map embeds into so that .
2. Mercer and Hermite Expansions: Explicit Feature Maps
On , Mercer’s theorem applies to the Gaussian kernel and gives an eigen-decomposition: The expansion is explicitly realized in terms of Hermite polynomials. For the one-dimensional case, with standard normal weight and orthonormal Hermite functions , the kernel admits
where the are scaled Hermite functions and depends explicitly on (Gnewuch et al., 2021). In higher dimensions, the eigenfunctions and eigenvalues tensorize across input coordinates.
The associated feature map is
where is the tensor product of univariate Hermite functions. For general (possibly infinite) , the maximal domain consists of sequences with , and the RKHS is the incomplete tensor product of the univariate spaces (Gnewuch et al., 2021).
3. Fourier (Bochner) Representations and Random Features
For shift-invariant kernels, Bochner’s theorem yields a spectral representation: For the Gaussian kernel, is a Gaussian measure with density (Jorgensen et al., 2017). The (complex-valued) feature map is and belongs to .
In practical applications, random Fourier features (RFF) approximate the Gaussian kernel by finite-dimensional real-valued feature maps (Ton et al., 2017). Drawing and ,
The inner product of finite feature vectors converges to at rate in spectral norm, as (Ton et al., 2017).
4. Polynomial and Maclaurin Expansions; Localized Feature Maps
Equivalently, the exponential can be expanded via Maclaurin series,
yielding the "polynomial sketch" feature map, where monomials are approximated by random projections (polynomial sketches) (Wacker et al., 2022). Localization—centering the feature map around the test point—cures the pathologies of naive Maclaurin truncation, yielding accurate, finite-dimensional local approximations, especially effective for high-frequency (short lengthscale) data regimes.
5. Finite-Dimensional Explicit Feature Maps for Finite Sets
For a finite dataset , an explicit, exact -dimensional feature map can be constructed (Ghiasi-Shirazi et al., 2024). Defining the kernel matrix , its inverse square root , and , one defines
This map reproduces the kernel exactly between any pair where at least one argument is a training point: . The explicit feature representation enables kernel PCA and other linear methods to be directly implemented in the primal space.
6. Fock and Segal–Bargmann Space Constructions
The Gaussian feature space can be realized as a Fock (Segal–Bargmann) space of entire functions ,
with an orthonormal basis of monomials and explicit reproducing kernel () (Alpay et al., 2022). The Segal–Bargmann transform provides an isometric isomorphism between and . The Gaussian kernel is realized as the inner product in this space,
with explicit Hermite-basis expansion.
7. Universality, ISPD, and Domain Extensions
The Gaussian kernel is universal, integrally strictly positive definite (ISPD), and -universal on Hilbert spaces; is dense in , ensuring that the feature-map embedding is rich enough to approximate all continuous functions vanishing at infinity (Guella, 2020). These properties ensure strong consistency and flexibility for statistical learning. Analogous constructions exist for Gaussian-type kernels on non-Euclidean domains, including hyperbolic spaces via Schoenberg’s theorem and generalizations to conditional negative definiteness (CND) kernels.
Table: Major Constructions of Gaussian Kernel Feature Spaces
| Construction Type | Space / Domain | Feature Map / Expansion |
|---|---|---|
| Hermite/Mercer Expansion | Hermite polynomials via Mercer, | |
| Bochner/Fourier Representation | (Fourier basis, RFF) | |
| Polynomial/Maclaurin | Monomials/polynomial sketches, localized | |
| Fock/Segal–Bargmann Space | Entire holomorphic functions, Hermite basis | |
| Explicit finite-dimensional | -point subset | -dimensional exact construction |
Each construction realizes the RKHS with different analytical and computational properties: Hermite gives orthonormal bases, Bochner yields spectral sampling, polynomial/Maclaurin facilitates sparse approximations, Fock spaces bring connections to complex analysis and quantum mechanics, while finite-dimensional explicit maps enable primal implementations over finite data sets.
References
- (Guella, 2020): On Gaussian kernels on Hilbert spaces and kernels on Hyperbolic spaces
- (Gnewuch et al., 2021): Countable Tensor Products of Hermite Spaces and Spaces of Gaussian Kernels
- (Jorgensen et al., 2017): Reproducing kernels and choices of associated feature spaces, in the form of -spaces
- (Wacker et al., 2022): Local Random Feature Approximations of the Gaussian Kernel
- (Ghiasi-Shirazi et al., 2024): An Exact Finite-dimensional Explicit Feature Map for Kernel Functions
- (Ton et al., 2017): Spatial Mapping with Gaussian Processes and Nonstationary Fourier Features
- (Alpay et al., 2022): An approach to the Gaussian RBF kernels via Fock spaces