Papers
Topics
Authors
Recent
Search
2000 character limit reached

Generalized Distance Covariance Overview

Updated 19 February 2026
  • Generalized distance covariance is a framework that measures dependence between random elements, enabling tests of independence under minimal moment conditions.
  • It employs negative definite functions derived from symmetric Lévy measures to generalize classical distance covariance while accommodating various distance structures.
  • Its versatile methodology unifies distance and Gaussian covariance concepts, offering consistent and efficient estimation for multivariate and functional data analysis.

Generalized distance covariance is a framework for measuring dependence between random elements—typically random vectors, but extendable to broad classes of metric or semimetric spaces—where traditional Euclidean distances and α-stable weights are replaced by negative definite functions derived from symmetric Lévy measures. This generalization preserves the core properties of distance covariance, such as characterizing independence and enabling efficient empirical estimation, while accommodating less restrictive moment conditions and a wide variety of alternative distance structures in finite or infinite dimensions. The most significant outcomes include broadened applicability, natural integration of process-based and metric-based dependence, explicit connections to Gaussian (Brownian) covariance, and foundational implications for distance multivariance and tests of mutual independence (Böttcher et al., 2017).

1. Formal Definition and General Framework

Let XRpX \in \mathbb{R}^p and YRqY \in \mathbb{R}^q be random vectors with joint characteristic function f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle} and marginals fX(s)f_X(s), fY(t)f_Y(t). The generalized distance covariance formally arises by selecting symmetric Lévy measures μ\mu (on Rp\mathbb{R}^p) and ν\nu (on Rq\mathbb{R}^q), forming the product measure ρ=μν\rho = \mu \otimes \nu on YRqY \in \mathbb{R}^q0, and constructing the associated real-valued continuous negative definite functions ("cndfs")

YRqY \in \mathbb{R}^q1

The generalized distance covariance is then

YRqY \in \mathbb{R}^q2

or, equivalently,

YRqY \in \mathbb{R}^q3

The classical Székely–Rizzo–Bakirov distance covariance is recovered as a special case by taking YRqY \in \mathbb{R}^q4 and YRqY \in \mathbb{R}^q5 to be (possibly isotropic) symmetric YRqY \in \mathbb{R}^q6-stable Lévy measures, in which case YRqY \in \mathbb{R}^q7 and YRqY \in \mathbb{R}^q8 for YRqY \in \mathbb{R}^q9 (Böttcher et al., 2017).

2. Theoretical Properties and Characterization of Independence

By construction, f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle}0 and

f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle}1

i.e., generalized distance covariance is zero if and only if f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle}2 and f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle}3 are independent. The moment conditions required for finiteness of f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle}4 relax those of the standard framework: it suffices that f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle}5 and f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle}6, compared to the original requirement that f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle}7 for f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle}8. Thus, substantially weaker moment assumptions are admissible in the Lévy-based generalization (Böttcher et al., 2017).

3. Alternative Distance Structures and Negative Definite Functions

By varying the underlying Lévy measures, one synthesizes alternative cndfs and “metric” structures compatible with generalized distance covariance. Examples include:

  • Minkowski f(X,Y)(s,t)=Eeis,X+it,Yf_{(X,Y)}(s,t) = \mathbb{E} e^{i\langle s,X\rangle + i\langle t,Y\rangle}9-distance for fX(s)f_X(s)0:

fX(s)f_X(s)1

arising from product measures of one-dimensional fX(s)f_X(s)2-stable Lévy laws.

  • Cauchy-type distance: fX(s)f_X(s)3, corresponding to isotropic Cauchy measures fX(s)f_X(s)4.
  • Truncated or compound Poisson distances: fX(s)f_X(s)5 for finite fX(s)f_X(s)6.

This flexibility enables practitioners to tailor the dependence measure to the specific probabilistic structure or application requirements, and was motivated explicitly by the aim of handling more general types of data and stochastic processes (Böttcher et al., 2017).

4. Computation, Empirical Estimation, and Consistency

Empirical generalized distance covariance is a V-statistic based on a four-fold kernel:

fX(s)f_X(s)7

The construction proceeds via distance matrices fX(s)f_X(s)8 and fX(s)f_X(s)9, which are double-centered to obtain fY(t)f_Y(t)0 and fY(t)f_Y(t)1 (using the centering matrix fY(t)f_Y(t)2). The estimator

fY(t)f_Y(t)3

is strongly consistent under minimal moment assumptions. When fY(t)f_Y(t)4, a central limit theorem holds for the scaled statistic fY(t)f_Y(t)5, whose limit is the squared fY(t)f_Y(t)6-norm of a centered Gaussian field corresponding to the chosen Lévy measure (Böttcher et al., 2017).

5. Relationship to Gaussian Covariance and Brownian Covariance

Generalized distance covariance unifies distance-based dependence with covariance structures defined via Gaussian fields. For a mean-zero Gaussian field fY(t)f_Y(t)7 (indexed over fY(t)f_Y(t)8) with covariance fY(t)f_Y(t)9, define the centered variables μ\mu0, and similarly for μ\mu1. Then,

μ\mu2

establishing exact agreement between generalized distance covariance and “Gaussian covariance,” with Brownian covariance corresponding to the special case μ\mu3 (Böttcher et al., 2017, Székely et al., 2010).

6. Extension to Multivariate, Functional, and General Metric Space Settings

Generalized distance covariance serves as the foundation for “distance multivariance,” a measure of mutual dependence among μ\mu4 random vectors, and has been characterized in a fully abstract metric space framework. Lyons (Lyons, 2011) demonstrated that the key requirement for the independence characterization is that metrics have strong negative type, encompassing all separable Hilbert spaces (and hence infinite-dimensional settings, functional data, and stochastic processes (Matsui et al., 2017, Janson, 2019)). The Hilbert space embedding connects generalized distance covariance to Hilbert–Schmidt operators, revealing that for μ\mu5, the measure specializes to four times the squared Hilbert–Schmidt norm of the cross-covariance operator (Janson, 2019).

7. Implications, Practical Use, and Limitations

Generalized distance covariance delivers a unified and versatile dependence measure with the following high-level features:

  • Precisely characterizes independence for a broad class of random elements under minimal moments.
  • Admits versions tailored to alternative distances, general kernels, or semimetrics (including kernel distance covariance and its Hilbert–Schmidt generalizations (Han et al., 2021)).
  • Enables efficient, consistent estimation and permutation-based inference.
  • Extends to mutual independence testing and independent component analysis (Jin et al., 2017, Matteson et al., 2013).
  • Proves critical in high-dimensional and functional data regimes, and demonstrates power universality and minimax optimality among tests relying on pairwise distances or kernels (Han et al., 2021). A key limitation is the dependence of definitions and equivalence of forms on moment conditions; the integral expressions may be infinite or undefined if the tail behavior of μ\mu6 and μ\mu7 is too heavy and required moments fail to exist (Janson, 2019). For practical implementation, practitioners must verify that underlying Lévy-based negative definite functions yield finite expectations.

References:

  • "Detecting independence of random vectors: generalized distance covariance and Gaussian covariance" (Böttcher et al., 2017)
  • "On distance covariance in metric and Hilbert spaces" (Janson, 2019)
  • "Distance covariance in metric spaces" (Lyons, 2011)
  • "Brownian distance covariance" (Székely et al., 2010)
  • "Generalized kernel distance covariance in high dimensions: non-null CLTs and power universality" (Han et al., 2021)
  • "Generalizing Distance Covariance to Measure and Test Multivariate Mutual Dependence" (Jin et al., 2017)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Generalized Distance Covariance.