Papers
Topics
Authors
Recent
Search
2000 character limit reached

Trace-Induced Quantum Kernels

Updated 21 February 2026
  • Trace-induced quantum kernels are similarity measures derived from quantum state embeddings using the trace of operator products, unifying global fidelity and local projected approaches.
  • Generalized kernels combine fundamental Lego kernels, allowing precise control over expressivity, inductive bias, and quantum resource trade-offs.
  • They underpin diverse applications from quantum time series analysis to multi-output regression, balancing performance with reduced quantum circuit demands.

A trace-induced quantum kernel is a family of similarity measures between inputs, constructed from quantum state embeddings where kernel values are given by the trace of operator products applied to density matrices. This framework encompasses global fidelity kernels, various local or projected quantum kernels, and operator-valued or entangled kernels. Trace-induced quantum kernels play a central role in quantum machine learning for encoding similarity in quantum feature spaces, with implications for expressivity, sample complexity, and quantum resource requirements (Gan et al., 2023, Shin et al., 26 Mar 2025, Huusari et al., 2021).

1. Mathematical Foundation and Definitions

Given a classical input xXx\in\mathcal{X}, a trace-induced quantum kernel is constructed by defining a data-embedding channel U(x)U(x) which generates an nn-qubit density operator: ρ(x)=U(x)ρ0U(x),\rho(x) = U(x)\rho_0 U(x)^\dagger, where ρ0\rho_0 is the initial state.

A trace-induced quantum kernel is defined as: k(x,x)=Tr[F(ρ(x))G(ρ(x))],k(x,x') = \mathrm{Tr}[ F(\rho(x))\, G(\rho(x')) ], where FF and GG are Hermitian-preserving linear superoperators. Significant special cases include:

  • Global Fidelity Quantum Kernel (GFQK):

kglobal(x,x)=Tr[ρ(x)ρ(x)],k_{\mathrm{global}}(x,x') = \mathrm{Tr}[\rho(x)\rho(x')],

which measures the Hilbert-Schmidt inner product (state overlap) of embedded density operators (Gan et al., 2023, Shin et al., 26 Mar 2025).

  • Local Projected Quantum Kernels (LPQK):

For a subset s{1,,n}s \subset \{1,\dots,n\}:

ks(x,x)=Trs[ρs(x)ρs(x)],withρs(x)=Trsˉ[ρ(x)].k_{s}(x,x') = \mathrm{Tr}_s [\rho_s(x)\rho_s(x')], \quad \text{with} \quad \rho_s(x) = \mathrm{Tr}_{\bar s}[\rho(x)].

Averaging over all subsets s=S|s|=S yields the SS-LPQK.

These constructions unify widely used quantum kernels and form the basis for systematic kernel design in quantum machine learning (Gan et al., 2023).

2. Generalization via Lego Kernels

All trace-induced quantum kernels can be expressed as (nonnegative) combinations of fundamental "Lego" kernels. Given an orthonormal operator basis {Ai}i=14n\{A_i\}_{i=1}^{4^n} (e.g., Pauli basis), a single Lego kernel is: ki(x,x)=Tr[ρ(x)Ai]Tr[ρ(x)Ai].k_i(x,x') = \mathrm{Tr}[\rho(x)A_i]\, \mathrm{Tr}[\rho(x')A_i]. A generalized trace-induced quantum kernel (GTQK) is a positive combination of Lego kernels: kW(x,x)=i=14nwiki(x,x)=i=14nwiTr[ρ(x)Ai]Tr[ρ(x)Ai],k_W(x,x') = \sum_{i=1}^{4^n} w_i\, k_i(x,x') = \sum_{i=1}^{4^n} w_i\, \mathrm{Tr}[\rho(x)A_i]\,\mathrm{Tr}[\rho(x')A_i], with wi0w_i\ge 0 and iwi2=1\sum_i w_i^2=1. All global and local quantum kernels correspond to specific choices of wiw_i (Gan et al., 2023):

  • GFQK: wi=1/2nw_i = 1/2^n (all features "on").
  • ss-LPQK: wi=1/2sw_i = 1/2^{|s|} if AiA_i is supported on ss, and zero otherwise.

A feature map perspective yields: kW(x,x)=ΦW(x),ΦW(x),    ΦW(x)=(wiTr[ρ(x)Ai])i=14n.k_W(x,x') = \langle \Phi_W(x), \Phi_W(x') \rangle, \;\; \Phi_W(x) = \left(\sqrt{w_i}\, \mathrm{Tr}[\rho(x) A_i ]\right)_{i=1}^{4^n}.

3. Expressivity, Inductive Bias, and Generalization

The representational power of a trace-induced quantum kernel is determined by the set and weights of the underlying Lego (operator) features. The associated RKHS is

HW={f()=i=14nβiwiTr[ρ()Ai]:iβi2<}.\mathcal{H}_W = \left\{ f(\cdot) = \sum_{i=1}^{4^n} \beta_i \sqrt{w_i}\, \mathrm{Tr}[\rho(\cdot)A_i] : \sum_i \beta_i^2 < \infty \right\}.

The number p=#{i:wi0}p = \#\{i: w_i \neq 0\} of active Lego features controls model complexity:

  • p=4np=4^n yields full GFQK expressivity.
  • Reduced pp (via sparse wiw_i) improves generalization by restricting capacity.

Under regularized empirical risk minimization, features with smaller wiw_i are penalized more. The empirical Rademacher complexity of the RKHS HW\mathcal{H}_W scales as O(p/N)O(\sqrt{\sqrt{p}/N}), so statistical complexity increases as p1/4p^{1/4} (Gan et al., 2023). This explicitly links inductive bias and generalization behavior to kernel structure.

4. Systematic Kernel Complexity and Resource Trade-offs

The trace-induced kernel formalism enables a stepwise approach to increasing kernel complexity, which directly correlates with quantum resource costs:

  • HH-body LPQKs: Restrict Lego kernels to Pauli operators of weight H\leq H:

k(H)(x,x)=i:wt(Pi)H1dHTr[ρ(x)Pi]Tr[ρ(x)Pi],  dH=h=0H(nh)3h.k^{(H)}(x,x') = \sum_{i:\, \mathrm{wt}(P_i)\leq H} \frac{1}{\sqrt{d_H}}\, \mathrm{Tr}[\rho(x)P_i]\, \mathrm{Tr}[\rho(x')P_i],\; d_H = \sum_{h=0}^{H} \binom{n}{h}3^h.

This construction systematically expands capacity while controlling measurement costs.

  • Resource scaling:
    • GFQK evaluation for NN data points requires O(N2)O(N^2) circuit executions, each of twice the embedding circuit depth.
    • HH-body LPQKs need only O(N)O(N) executions, with circuit depth identical to U(x)U(x), since all required observables can be measured via classical shadows or similar protocols.

Empirical results on benchmark datasets demonstrate that 2-body LPQKs with p150p\sim 150 features can match GFQK classification accuracy while requiring orders of magnitude fewer gate applications and measurement shots (Gan et al., 2023).

5. Entangled, Operator-Valued, and Tensor-Kernel Perspectives

Trace-induced quantum kernels are a specific instance of the broader class of entangled tensor kernels (ETKs) (Shin et al., 26 Mar 2025). An ETK has the form

KC(x,y)=F(x)CF(y),K_C(x,y) = \langle F(x) | C | F(y) \rangle,

where F(x)F(x) is a tensor product of local feature maps, and CC is a positive semidefinite "core" matrix.

All embedding quantum kernels with fidelity structure,

K(x,y)=Tr[ρ(x)ρ(y)],K(x,y) = \mathrm{Tr}[\rho(x)\rho(y)],

can be written as ETKs, where F(x)F(x) represents a trigonometric feature (e.g., k[1,cosϕk(x),sinϕk(x)]\otimes_{k} [1, \cos\, \phi_k(x), \sin\, \phi_k(x)]), and CTC_T encodes the data-independent circuit structure. The entangling structure of CTC_T introduces global correlations not present in simple product kernels, which governs inductive bias, expressivity (Fourier spectrum), and complexity.

In the context of operator-valued kernels (Huusari et al., 2021), the partial trace construction yields matrix-valued kernels encoding output structure or input-output "entanglement". The Choi–Kraus decomposition translates these into finite-rank operator sums, compatible with classical kernel alignment and regression pipelines.

6. Applications in Time Series and Generative Models

Trace-induced quantum kernels are applied beyond standard supervised learning tasks:

  • Quantum Time Series Kernels: When data is generated by a quantum hidden Markov model (QHMM), a trace-based kernel can be induced by mapping sequences xx to quantum states ρx\rho_x, then defining similarity via a function of the trace-norm distance:

K(x,y)=exp(γdtr(ρx,ρy)2),    dtr(ρ,σ)=12ρσ1,K(x,y) = \exp\left( -\gamma\, d_{\mathrm{tr}}(\rho_x, \rho_y)^2 \right),\;\; d_{\mathrm{tr}}(\rho, \sigma) = \tfrac12 \|\rho - \sigma\|_1,

which is guaranteed positive semidefinite (Mercer kernel) (Markov et al., 2023). This approach outperforms classical RBF kernels in experiments on sequence classification tasks, with improved separation as the Hilbert space dimension increases.

  • Multivariate Outputs / Multi-output Regression: Operator-valued, trace-induced kernels enable modeling input-output correlations in problems with structured or high-dimensional labels. Entangled kernels learned via Riemannian optimization outperform ordinary kernel ridge regression and input-output separable kernels, especially when the number of outputs approaches or exceeds the sample size (Huusari et al., 2021).

7. Implementation Guidelines and Dequantization Strategies

Practical implementation of trace-induced quantum kernels involves choices and trade-offs:

  • Select a problem-tailored circuit ansatz U(x)U(x) and kernel bandwidth (data scaling).
  • Start with low-weight (HH-body) LPQKs to limit quantum resource requirements, and increase complexity if performance gains justify it.
  • Use classical shadows or other efficient tomography techniques to estimate requisite observable expectation values.
  • Regularize kernel complexity and model parameters (e.g., SVM penalty) via cross-validation.
  • If lower-body LPQKs achieve comparable performance to GFQK on the target task, this yields significant resource savings.

For dequantization and classical simulation, one may approximate the entangled core tensor CTC_T using low-bond-dimension matrix product operator (MPO) techniques, learn surrogate ETKs, or use randomized (Fourier) feature strategies—each exploiting the insight that kernel expressivity and computational hardness are encoded in the spectrum and structure of CTC_T (Shin et al., 26 Mar 2025).


Key references:

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Trace-Induced Quantum Kernels.