Distance correlation
In statistics and in probability theory, distance correlation is a measure of statistical dependence between two random variables or two random vectors of arbitrary, not necessarily equal dimension. Its important property is that this measure of dependence is zero if and only if the random variables are statistically independent. This measure is derived from a number of other quantities that are used in its specification, specifically: distance variance, distance standard deviation and distance covariance. These take the same roles as the ordinary moments with corresponding names in the specification of the Pearson product-moment correlation coefficient.
These distance-based measures can be put into an indirect relationship to the ordinary moments by an alternative formulation (described below) using ideas related to Brownian motion, and this has led to the use of names such as Brownian covariance and Brownian distance covariance.
Background
The classical measure of dependence, the Pearson correlation coefficient,[1] is mainly sensitive to a linear relationship between two variables. Distance correlation was introduced in 2005 by Gabor J Szekely in several lectures to address this deficiency of Pearson’s correlation, namely that it can easily be zero for dependent variables. Correlation = 0 (uncorrelatedness) does not imply independence while distance correlation = 0 does imply independence. The first results on distance correlation were published in 2007 and 2009.[2][3] It was proved that distance covariance is the same as the Brownian covariance.[3] These measures are examples of energy distances.
Definitions
Distance covariance
The population value of distance covariance [2]: p.2783, [4] is the square root of
where denotes expected value, denotes Euclidean norm, and , , and are independent and identically distributed. Distance covariance can be expressed in terms of Pearson’s covariance, , as follows: . This identity shows that the distance covariance is not the same as the covariance of distances, , which can be zero even if and are not independent.
The sample distance covariance is defined as follows. Let , be a statistical sample from a pair of real valued or vector valued random variables . First, compute all pairwise distances
- and for .
That is, compute the by distance matrices and . Then take all centered distances and where is the -th row mean, is the -th column mean, and is the grand mean of the distance matrix of the sample. The notation is similar for the values. (In the matrices of centered distances and all row sums and all column sums equal zero.) The squared sample distance covariance is simply the arithmetic average of the products ; that is
The statistic determines a consistent multivariate test of independence of random vectors in arbitrary dimensions. For an implementation see dcov.test function in the energy package for R.[5]
Distance variance
The distance variance is a special case of distance covariance when the two variables are identical. The population value of distance variance is the square root of
where denotes the expected value, is an independent and identically distributed copy of and is independent of and and has the same distribution as and .
The sample distance variance is the square root of
which is a relative of Corrado Gini’s mean difference introduced in 1912 (but Gini did not work with centered distances).
Distance standard deviation
The distance standard deviation is the square root of the distance variance.
Distance correlation
The distance correlation [2][3] of two random variables is obtained by dividing their distance covariance by the product of their distance standard deviations. The distance correlation is
and the sample distance correlation is defined by substituting the sample distance covariance and distance variances for the population coefficients above.
For easy computation of sample distance correlation see the dcor function in the energy package for R.[5]
Properties
Distance correlation
(i) and .
(ii) if and only if and are independent.
(iii) implies that dimensions of the linear spaces spanned by and samples respectively are almost surely equal and if we assume that these subspaces are equal, then in this subspace for some vector , scalar , and orthonormal matrix .
Distance covariance
(i) and .
(ii) for all constant vectors , scalars , and orthonormal matrices .
(iii) If the random vectors and are independent then
Equality holds if and only if and are both constants, or and are both constants, or are mutually independent.
(iv) if and only if and are independent.
This last property is the most important effect of working with centered distances.
The statistic is a biased estimator of due to
The bias therefore can easily be corrected.[6]
Distance variance
(i) if and only if almost surely.
(ii) if and only if every sample observation is identical.
(iii) for all constant vectors , scalars , and orthonormal matrices .
(iv) If and are independent then .
Equality holds if (iv) if and only if one of the random variables or is a constant.
Generalization
Distance covariance can be generalized to include powers of Euclidean distance. Define
Then for every , and are independent if and only if . It is important to note that this characterization does not hold for exponent ; in this case for bivariate , is a deterministic function of the Pearson correlation.[2] If and are powers of the corresponding distances, , then sample distance covariance can be defined as the nonnegative number for which
One can extend to metric-space-valued random variables and : If has law in a metric space with metric , then define , , and (provided is finite, i.e., has finite first moment), . Then if has law (in a possibly different metric space with finite first moment), define
This is non-negative for all such iff both metric spaces have negative type.[7] Here, a metric space has negative type if is isometric to a subset of a Hilbert space[8]. If both metric spaces have strong negative type, then iff are independent.[7]
Alternative formulation: Brownian covariance
Brownian covariance is motivated by generalization of the notion of covariance to stochastic processes. The square of the covariance of random variables X and Y can be written in the following form:
where E denotes the expected value and the prime denotes independent and identically distributed copies. We need the following generalization of this formula. If U(s), V(t) are arbitrary random processes defined for all real s and t then define the U-centered version of X by
whenever the subtracted conditional expected value exists and denote by YV the V-centered version of Y.[3][9][10] The (U,V) covariance of (X,Y) is defined as the nonnegative number whose square is
whenever the right-hand side is nonnegative and finite. The most important example is when U and V are two-sided independent Brownian motions /Wiener processes with expectation zero and covariance |s| + |t| - |s-t| = 2 min(s,t). (This is twice the covariance of the standard Wiener process; here the factor 2 simplifies the computations.) In this case the (U,V) covariance is called Brownian covariance and is denoted by
There is a surprising coincidence: The Brownian covariance is the same as the distance covariance:
and thus Brownian correlation is the same as distance correlation.
On the other hand, if we replace the Brownian motion with the deterministic identity function id then Covid(X,Y) is simply the absolute value of the classical Pearson covariance,
See also
- RV coefficient
- For a related third-order statistic, see Distance skewness.
Notes
- ^ Pearson (1895)
- ^ a b c d Székely, Rizzo and Bakirov (2007) Cite error: The named reference "SR2007" was defined multiple times with different content (see the help page).
- ^ a b c d Székely & Rizzo (2009)
- ^ Székely & Rizzo (2009) Theorem 7, (3.7), p. 1249.
- ^ a b energy package for R
- ^ Székely and Rizzo (2009), Rejoinder
- ^ a b Lyons, R. (2011) Distance covariance in metric spaces [1]
- ^ Klebanov, L. B. (2005) N-distances and their Applications, Karolinum Press, Charles University, Prague.
- ^ Bickel & Xu (2009)
- ^ Kosorok (2009)
References
- Bickel, P.J. and Xu, Y. (2009) "Discussion of: Brownian distance covariance", Annals of Applied Statistics, 3 (4), 1266–1269. doi:10.1214/09-AOAS312A Free access to article
- Gini, C. (1912). Variabilità e Mutabilità. Bologna: Tipografia di Paolo Cuppini.
- Pearson, K. (1895). "Note on regression and inheritance in the case of two parents", Proceedings of the Royal Society, 58, 240–242
- Pearson, K. (1920). "Notes on the history of correlation", Biometrika, 13, 25–45.
- Székely, G. J. Rizzo, M. L. and Bakirov, N. K. (2007). "Measuring and testing independence by correlation of distances", Annals of Statistics, 35/6, 2769–2794. doi:10.1214/009053607000000505 Reprint
- Székely, G. J. and Rizzo, M. L. (2009). "Brownian distance covariance", Annals of Applied Statistics, 3/4, 1233–1303. doi:10.1214/09-AOAS312 Reprint
- Kosorok, M. R. (2009) "Discussion of: Brownian Distance Covariance", Annals of Applied Statistics, 3/4, 1270–1278. doi:10.1214/09-AOAS312B Free access to article