Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Kendall rank correlation coefficient

From Wikipedia, the free encyclopedia
Statistic for rank correlation
"Tau-a" redirects here. For the astronomical radio source, seeTaurus A.
"Tau coefficient" redirects here; not to be confused withTau distribution.

Instatistics, theKendall rank correlation coefficient, commonly referred to asKendall's τ coefficient (after the Greek letterτ, tau), is astatistic used to measure theordinal association between two measured quantities. Aτ test is anon-parametrichypothesis test for statistical dependence based on the τ coefficient. It is a measure ofrank correlation: the similarity of the orderings of the data whenranked by each of the quantities. It is named afterMaurice Kendall, who developed it in 1938,[1] thoughGustav Fechner had proposed a similar measure in the context oftime series in 1897.[2]

Intuitively, the Kendall correlation between two variables will be high when observations have a similar (or identical for a correlation of 1)rank (i.e. relative position label of the observations within the variable: 1st, 2nd, 3rd, etc.) between the two variables, and low when observations have a dissimilar (or fully different for a correlation of −1) rank between the two variables.

Both Kendall'sτ{\displaystyle \tau } andSpearman'sρ{\displaystyle \rho } can be formulated as special cases of a moregeneral correlation coefficient. Its notions ofconcordance and discordance also appear in other areas of statistics, like theRand index incluster analysis.

Definition

[edit]
All points in the gray area are concordant and all points in the white area are discordant with respect to point(X1,Y1){\displaystyle (X_{1},Y_{1})}. Withn=30{\displaystyle n=30} points, there are a total of(302)=435{\displaystyle {\binom {30}{2}}=435} possible point pairs. In this example there are 395 concordant point pairs and 40 discordant point pairs, leading to a Kendall rank correlation coefficient of 0.816.

Let(x1,y1),...,(xn,yn){\displaystyle (x_{1},y_{1}),...,(x_{n},y_{n})} be a set of observations of the joint random variablesX andY, such that all the values of (xi{\displaystyle x_{i}}) and (yi{\displaystyle y_{i}}) are unique. (See the section#Accounting for ties for ways of handling non-unique values.) Any pair of observations(xi,yi){\displaystyle (x_{i},y_{i})} and(xj,yj){\displaystyle (x_{j},y_{j})}, wherei<j{\displaystyle i<j}, are said to beconcordant if the sort order of(xi,xj){\displaystyle (x_{i},x_{j})} and(yi,yj){\displaystyle (y_{i},y_{j})} agrees: that is, if either bothxi>xj{\displaystyle x_{i}>x_{j}} andyi>yj{\displaystyle y_{i}>y_{j}} holds or bothxi<xj{\displaystyle x_{i}<x_{j}} andyi<yj{\displaystyle y_{i}<y_{j}}; otherwise they are said to bediscordant.

In the absence of ties, the Kendall τ coefficient is defined as:

τ=(number of concordant pairs)(number of discordant pairs)(number of pairs)=12(number of discordant pairs)(n2).{\displaystyle \tau ={\frac {({\text{number of concordant pairs}})-({\text{number of discordant pairs}})}{({\text{number of pairs}})}}=1-{\frac {2({\text{number of discordant pairs}})}{n \choose 2}}.}[3]

fori<j<n{\displaystyle i<j<n} where(n2)=n(n1)2{\displaystyle {n \choose 2}={n(n-1) \over 2}} is thebinomial coefficient for the number of ways to choose two items from n items.

The number of discordant pairs is equal to theinversion number that permutes the y-sequence into the same order as the x-sequence.

Properties

[edit]

Thedenominator is the total number of pair combinations, so the coefficient must be in the range −1 ≤ τ ≤ 1.

Hypothesis test

[edit]

The Kendall rank coefficient is often used as atest statistic in astatistical hypothesis test to establish whether two variables may be regarded as statistically dependent. This test isnon-parametric, as it does not rely on any assumptions on the distributions ofX orY or the distribution of (X,Y).

Under thenull hypothesis of independence ofX andY, thesampling distribution ofτ has anexpected value of zero. The precise distribution cannot be characterized in terms of common distributions, but may be calculated exactly for small samples; for larger samples, it is common to use an approximation to thenormal distribution, with mean zero and variance2(2n+5)/9n(n1){\textstyle 2(2n+5)/9n(n-1)}.[4]

Theorem. If the samples are independent, then the variance ofτA{\textstyle \tau _{A}} is given byVar[τA]=2(2n+5)/9n(n1){\textstyle Var[\tau _{A}]=2(2n+5)/9n(n-1)}.

Proof
Proof
Valz & McLeod (1990;[5] 1995[6])

WLOG, we reorder the data pairs, so thatx1<x2<<xn{\textstyle x_{1}<x_{2}<\cdots <x_{n}}. By assumption of independence, the order ofy1,...,yn{\textstyle y_{1},...,y_{n}} is a permutation sampled uniformly at random fromSn{\textstyle S_{n}}, the permutation group on1:n{\textstyle 1:n}.

For each permutation, its uniquel{\textstyle l}inversion code isl0l1ln1{\textstyle l_{0}l_{1}\cdots l_{n-1}} such that eachli{\textstyle l_{i}} is in the range0:i{\textstyle 0:i}. Sampling a permutation uniformly is equivalent to sampling al{\textstyle l}-inversion code uniformly, which is equivalent to sampling eachli{\textstyle l_{i}} uniformly and independently.

Then we haveE[τA2]=E[(14ilin(n1))2]=18n(n1)iE[li]+16n2(n1)2ijE[lilj]=18n(n1)iE[li]+16n2(n1)2(ijE[li]E[lj]+iV[li])=18n(n1)iE[li]+16n2(n1)2ijE[li]E[lj]+16n2(n1)2(iV[li])=(14iE[li]n(n1))2+16n2(n1)2(iV[li]){\displaystyle {\begin{aligned}E[\tau _{A}^{2}]&=E\left[\left(1-{\frac {4\sum _{i}l_{i}}{n(n-1)}}\right)^{2}\right]\\&=1-{\frac {8}{n(n-1)}}\sum _{i}E[l_{i}]+{\frac {16}{n^{2}(n-1)^{2}}}\sum _{ij}E[l_{i}l_{j}]\\&=1-{\frac {8}{n(n-1)}}\sum _{i}E[l_{i}]+{\frac {16}{n^{2}(n-1)^{2}}}\left(\sum _{ij}E[l_{i}]E[l_{j}]+\sum _{i}V[l_{i}]\right)\\&=1-{\frac {8}{n(n-1)}}\sum _{i}E[l_{i}]+{\frac {16}{n^{2}(n-1)^{2}}}\sum _{ij}E[l_{i}]E[l_{j}]+{\frac {16}{n^{2}(n-1)^{2}}}\left(\sum _{i}V[l_{i}]\right)\\&=\left(1-{\frac {4\sum _{i}E[l_{i}]}{n(n-1)}}\right)^{2}+{\frac {16}{n^{2}(n-1)^{2}}}\left(\sum _{i}V[l_{i}]\right)\end{aligned}}}

The first term is justE[τA]2=0{\textstyle E[\tau _{A}]^{2}=0}. The second term can be calculated by noting thatli{\textstyle l_{i}} is a uniform random variable on0:i{\textstyle 0:i}, soE[li]=i2{\textstyle E[l_{i}]={\frac {i}{2}}} andE[li2]=02++i2i+1=i(2i+1)6{\textstyle E[l_{i}^{2}]={\frac {0^{2}+\cdots +i^{2}}{i+1}}={\frac {i(2i+1)}{6}}}, then using the sum of squares formula again.

Asymptotic normalityAt then{\textstyle n\to \infty } limit,zA=τAVar[τA]=nCnDn(n1)(2n+5)/18{\textstyle z_{A}={\frac {\tau _{A}}{\sqrt {Var[\tau _{A}]}}}={n_{C}-n_{D} \over {\sqrt {n(n-1)(2n+5)/18}}}} converges in distribution to the standard normal distribution.

Proof

Use a result fromA class of statistics with asymptotically normal distribution Hoeffding (1948).[7]

Case of standard normal distributions

[edit]

If(x1,y1),(x2,y2),...,(xn,yn){\textstyle (x_{1},y_{1}),(x_{2},y_{2}),...,(x_{n},y_{n})} are IID samples from the same jointly normal distribution with a knownPearson correlation coefficientr{\textstyle r}, then the expectation of Kendall rank correlation has a closed-form formula.[8]

Greiner's equalityIfX,Y{\textstyle X,Y} are jointly normal, with correlationr{\textstyle r}, thenr=sin(π2E[τA]){\displaystyle r=\sin {\left({\frac {\pi }{2}}E[\tau _{A}]\right)}}

The name is credited to Richard Greiner (1909)[9] byP. A. P. Moran.[10]

Proof
Proof[11]

Define the following quantities.

In the notation, we see that the number of concordant pairs,nC{\textstyle n_{C}}, is equal to the number ofΔi,j{\textstyle \Delta _{i,j}} that fall in the subsetA+{\textstyle A^{+}}. That is,nC=1i<jn1Δi,jA+{\textstyle n_{C}=\sum _{1\leq i<j\leq n}1_{\Delta _{i,j}\in A^{+}}}.

Thus,E[τA]=4n(n1)E[nC]1=4n(n1)1i<jnPr(Δi,jA+)1{\displaystyle E[\tau _{A}]={\frac {4}{n(n-1)}}E[n_{C}]-1={\frac {4}{n(n-1)}}\sum _{1\leq i<j\leq n}Pr(\Delta _{i,j}\in A^{+})-1}

Since each(xi,yi){\textstyle (x_{i},y_{i})} is an IID sample of the jointly normal distribution, the pairing does not matter, so each term in the summation is exactly the same, and soE[τA]=2Pr(Δ1,2A+)1{\displaystyle E[\tau _{A}]=2Pr(\Delta _{1,2}\in A^{+})-1} and it remains to calculate the probability. We perform this by repeated affine transforms.

First normalizeX,Y{\textstyle X,Y} by subtracting the mean and dividing the standard deviation. This does not changeτA{\textstyle \tau _{A}}. This gives us[xy]=[1rr1]1/2[zw]{\displaystyle {\begin{bmatrix}x\\y\end{bmatrix}}={\begin{bmatrix}1&r\\r&1\end{bmatrix}}^{1/2}{\begin{bmatrix}z\\w\end{bmatrix}}} where(Z,W){\textstyle (Z,W)} is sampled from the standard normal distribution onR2{\textstyle \mathbb {R} ^{2}}.

Thus,Δ1,2=2[1rr1]1/2[(z1z2)/2(w1w2)/2]{\displaystyle \Delta _{1,2}={\sqrt {2}}{\begin{bmatrix}1&r\\r&1\end{bmatrix}}^{1/2}{\begin{bmatrix}(z_{1}-z_{2})/{\sqrt {2}}\\(w_{1}-w_{2})/{\sqrt {2}}\end{bmatrix}}} where the vector[(z1z2)/2(w1w2)/2]{\textstyle {\begin{bmatrix}(z_{1}-z_{2})/{\sqrt {2}}\\(w_{1}-w_{2})/{\sqrt {2}}\end{bmatrix}}} is still distributed as the standard normal distribution onR2{\textstyle \mathbb {R} ^{2}}. It remains to perform some unenlightening tedious matrix exponentiations and trigonometry, which can be skipped over.

Thus,Δ1,2A+{\textstyle \Delta _{1,2}\in A^{+}} iff[(z1z2)/2(w1w2)/2]12[1rr1]1/2A+=122[11+r+11r11+r11r11+r11r11+r+11r]A+{\displaystyle {\begin{bmatrix}(z_{1}-z_{2})/{\sqrt {2}}\\(w_{1}-w_{2})/{\sqrt {2}}\end{bmatrix}}\in {\frac {1}{\sqrt {2}}}{\begin{bmatrix}1&r\\r&1\end{bmatrix}}^{-1/2}A^{+}={\frac {1}{2{\sqrt {2}}}}{\begin{bmatrix}{\frac {1}{\sqrt {1+r}}}+{\frac {1}{\sqrt {1-r}}}&{\frac {1}{\sqrt {1+r}}}-{\frac {1}{\sqrt {1-r}}}\\{\frac {1}{\sqrt {1+r}}}-{\frac {1}{\sqrt {1-r}}}&{\frac {1}{\sqrt {1+r}}}+{\frac {1}{\sqrt {1-r}}}\end{bmatrix}}A^{+}} where the subset on the right is a “squashed” version of two quadrants. Since the standard normal distribution is rotationally symmetric, we need only calculate the angle spanned by each squashed quadrant.

The first quadrant is the sector bounded by the two rays(1,0),(0,1){\textstyle (1,0),(0,1)}. It is transformed to the sector bounded by the two rays(11+r+11r,11+r11r){\textstyle ({\frac {1}{\sqrt {1+r}}}+{\frac {1}{\sqrt {1-r}}},{\frac {1}{\sqrt {1+r}}}-{\frac {1}{\sqrt {1-r}}})} and(11+r11r,11+r+11r){\textstyle ({\frac {1}{\sqrt {1+r}}}-{\frac {1}{\sqrt {1-r}}},{\frac {1}{\sqrt {1+r}}}+{\frac {1}{\sqrt {1-r}}})}. They respectively make angleθ{\textstyle \theta } with the horizontal and vertical axis, whereθ=arctan11+r11r11+r+11r{\displaystyle \theta =\arctan {\frac {{\frac {1}{\sqrt {1+r}}}-{\frac {1}{\sqrt {1-r}}}}{{\frac {1}{\sqrt {1+r}}}+{\frac {1}{\sqrt {1-r}}}}}}

Together, the two transformed quadrants span an angle ofπ+4θ{\textstyle \pi +4\theta }, soPr(Δ1,2A+)=π+4θ2π{\displaystyle Pr(\Delta _{1,2}\in A^{+})={\frac {\pi +4\theta }{2\pi }}} and therefore
sin(π2E[τA])=sin(2θ)=r{\displaystyle \sin {\left({\frac {\pi }{2}}E[\tau _{A}]\right)}=\sin(2\theta )=r}

Accounting for ties

[edit]

A pair{(xi,yi),(xj,yj)}{\displaystyle \{(x_{i},y_{i}),(x_{j},y_{j})\}} is said to betied if and only ifxi=xj{\displaystyle x_{i}=x_{j}} oryi=yj{\displaystyle y_{i}=y_{j}}; a tied pair is neither concordant nor discordant. When tied pairs arise in the data, the coefficient may be modified in a number of ways to keep it in the range [−1, 1]:

Tau-a

[edit]

The Tau statistic defined by Kendall in 1938[1] was retrospectively renamed Tau-a. It represents the strength of positive or negative association of two quantitative orordinal variables without any adjustment for ties. It is defined as:

τA=ncndn0{\displaystyle \tau _{A}={\frac {n_{c}-n_{d}}{n_{0}}}}

wherenc,nd andn0 are defined as in the next section.

When ties are present,nc+nd<n0{\displaystyle n_{c}+n_{d}<n_{0}} and, the coefficient can never be equal to +1 or -1. Even a perfect equality of the two variables (X=Y) leads to a Tau-a < 1.

Tau-b

[edit]

The Tau-b statistic, unlike Tau-a, makes adjustments for ties.This Tau-b was first described by Kendall in 1945 under the name Tau-w[12] as an extension of the original Tau statistic supporting ties.Values of Tau-b range from −1 (100% negative association, or perfect disagreement) to +1 (100% positive association, or perfect agreement). In case of the absence of association, Tau-b is equal to zero.

The Kendall Tau-b coefficient is defined as :

τB=ncnd(n0n1)(n0n2){\displaystyle \tau _{B}={\frac {n_{c}-n_{d}}{\sqrt {(n_{0}-n_{1})(n_{0}-n_{2})}}}}

where

n0=n(n1)/2n1=iti(ti1)/2n2=juj(uj1)/2nc=Number of concordant pairs, i.e. pairs with 0<i<j<n where xi<xj and yi<yj or xi>xj and yi>yjnd=Number of discordant, i.e. pairs where 0<i<j<n where xi<xj and yi>yj or xi<xj and yi>yjti=Number of tied values in the ith group of ties for the empirical distribution of Xuj=Number of tied values in the jth group of ties for the empirical distribution of Y{\displaystyle {\begin{aligned}n_{0}&=n(n-1)/2\\n_{1}&=\sum _{i}t_{i}(t_{i}-1)/2\\n_{2}&=\sum _{j}u_{j}(u_{j}-1)/2\\n_{c}&={\text{Number of concordant pairs, i.e. pairs with }}0<i<j<n{\text{ where }}x_{i}<x_{j}{\text{ and }}y_{i}<y_{j}{\text{ or }}x_{i}>x_{j}{\text{ and }}y_{i}>y_{j}\\n_{d}&={\text{Number of discordant, i.e. pairs where }}0<i<j<n{\text{ where }}x_{i}<x_{j}{\text{ and }}y_{i}>y_{j}{\text{ or }}x_{i}<x_{j}{\text{ and }}y_{i}>y_{j}\\t_{i}&={\text{Number of tied values in the }}i^{\text{th}}{\text{ group of ties for the empirical distribution of X}}\\u_{j}&={\text{Number of tied values in the }}j^{\text{th}}{\text{ group of ties for the empirical distribution of Y}}\end{aligned}}}

A simple algorithm developed in BASIC computes Tau-b coefficient using an alternative formula.[13]

Be aware that some statistical packages, e.g. SPSS, use alternative formulas for computational efficiency, with double the 'usual' number of concordant and discordant pairs.[14]

Tau-c

[edit]

Tau-c (also called Stuart-Kendall Tau-c)[15] was first defined by Stuart in 1953.[16]Contrary to Tau-b, Tau-c can be equal to +1 or -1 for non-square (i.e. rectangular)contingency tables,[15][16] i.e. when the underlying scale of both variables have different number of possible values. For instance, if the variable X has a continuous uniform distribution between 0 and 100 and Y is a dichotomous variable equal to 1 if X ≥ 50 and 0 if X < 50, the Tau-c statistic of X and Y is equal to 1 while Tau-b is equal to 0.707. A Tau-C equal to 1 can be interpreted as the best possible positive correlation conditional to marginal distributions while a Tau-B equal to 1 can be interpreted as the perfect positive monotonic correlation where the distribution of X conditional to Y has zero variance and the distribution of Y conditional to X has zero variance so that a bijective function f with f(X)=Y exists.

The Stuart-Kendall Tau-c coefficient is defined as:[16]

τC=2(ncnd)n2(m1)m=τAn1nmm1{\displaystyle \tau _{C}={\frac {2(n_{c}-n_{d})}{n^{2}{\frac {(m-1)}{m}}}}=\tau _{A}{\frac {n-1}{n}}{\frac {m}{m-1}}}

where

nc=Number of concordant pairsnd=Number of discordant pairsr=Number of rows of the contingency table (i.e. number of distinct xi)c=Number of columns of the contingency table (i.e. number of distinct yi)m=min(r,c){\displaystyle {\begin{aligned}n_{c}&={\text{Number of concordant pairs}}\\n_{d}&={\text{Number of discordant pairs}}\\r&={\text{Number of rows of the contingency table (i.e. number of distinct }}x_{i}{\text{)}}\\c&={\text{Number of columns of the contingency table (i.e. number of distinct }}y_{i}{\text{)}}\\m&=\min(r,c)\end{aligned}}}

Significance tests

[edit]

When two quantities are statistically dependent, the distribution ofτ{\displaystyle \tau } is not easily characterizable in terms of known distributions. However, forτA{\displaystyle \tau _{A}} the following statistic,zA{\displaystyle z_{A}}, is approximately distributed as a standard normal when the variables are statistically independent:

zA=ncnd118v0{\displaystyle z_{A}={n_{c}-n_{d} \over {\sqrt {{\frac {1}{18}}v_{0}}}}}

wherev0=n(n1)(2n+5){\displaystyle v_{0}=n(n-1)(2n+5)}.

Thus, to test whether two variables are statistically dependent, one computeszA{\displaystyle z_{A}}, and finds the cumulative probability for a standard normal distribution at|zA|{\displaystyle -|z_{A}|}. For a 2-tailed test, multiply that number by two to obtain thep-value. If thep-value is below a given significance level, one rejects the null hypothesis (at that significance level) that the quantities are statistically independent.

Numerous adjustments should be added tozA{\displaystyle z_{A}} when accounting for ties. The following statistic,zB{\displaystyle z_{B}}, has the same distribution as theτB{\displaystyle \tau _{B}} distribution, and is again approximately equal to a standard normal distribution when the quantities are statistically independent:

zB=ncndv{\displaystyle z_{B}={n_{c}-n_{d} \over {\sqrt {v}}}}

where

v=118v0(vt+vu)/18+(v1+v2)v0=n(n1)(2n+5)vt=iti(ti1)(2ti+5)vu=juj(uj1)(2uj+5)v1=iti(ti1)juj(uj1)/(2n(n1))v2=iti(ti1)(ti2)juj(uj1)(uj2)/(9n(n1)(n2)){\displaystyle {\begin{array}{ccl}v&=&{\frac {1}{18}}v_{0}-(v_{t}+v_{u})/18+(v_{1}+v_{2})\\v_{0}&=&n(n-1)(2n+5)\\v_{t}&=&\sum _{i}t_{i}(t_{i}-1)(2t_{i}+5)\\v_{u}&=&\sum _{j}u_{j}(u_{j}-1)(2u_{j}+5)\\v_{1}&=&\sum _{i}t_{i}(t_{i}-1)\sum _{j}u_{j}(u_{j}-1)/(2n(n-1))\\v_{2}&=&\sum _{i}t_{i}(t_{i}-1)(t_{i}-2)\sum _{j}u_{j}(u_{j}-1)(u_{j}-2)/(9n(n-1)(n-2))\end{array}}}

This is sometimes referred to as the Mann-Kendall test.[17]

Algorithms

[edit]

The direct computation of the numeratorncnd{\displaystyle n_{c}-n_{d}}, involves two nested iterations, as characterized by the following pseudocode:

numer := 0for i := 2..Ndofor j := 1..(i − 1)do        numer := numer + sign(x[i] − x[j]) × sign(y[i] − y[j])return numer

Although quick to implement, this algorithm isO(n2){\displaystyle O(n^{2})} in complexity and becomes very slow on large samples. A more sophisticated algorithm[18] built upon theMerge Sort algorithm can be used to compute the numerator inO(nlogn){\displaystyle O(n\cdot \log {n})} time.

Begin by ordering your data points sorting by the first quantity,x{\displaystyle x}, and secondarily (among ties inx{\displaystyle x}) by the second quantity,y{\displaystyle y}. With this initial ordering,y{\displaystyle y} is not sorted, and the core of the algorithm consists of computing how many steps aBubble Sort would take to sort this initialy{\displaystyle y}. An enhancedMerge Sort algorithm, withO(nlogn){\displaystyle O(n\log n)} complexity, can be applied to compute the number of swaps,S(y){\displaystyle S(y)}, that would be required by aBubble Sort to sortyi{\displaystyle y_{i}}. Then the numerator forτ{\displaystyle \tau } is computed as:

ncnd=n0n1n2+n32S(y),{\displaystyle n_{c}-n_{d}=n_{0}-n_{1}-n_{2}+n_{3}-2S(y),}

wheren3{\displaystyle n_{3}} is computed liken1{\displaystyle n_{1}} andn2{\displaystyle n_{2}}, but with respect to the joint ties inx{\displaystyle x} andy{\displaystyle y}.

AMerge Sort partitions the data to be sorted,y{\displaystyle y} into two roughly equal halves,yleft{\displaystyle y_{\mathrm {left} }} andyright{\displaystyle y_{\mathrm {right} }}, then sorts each half recursive, and then merges the two sorted halves into a fully sorted vector. The number ofBubble Sort swaps is equal to:

S(y)=S(yleft)+S(yright)+M(Yleft,Yright){\displaystyle S(y)=S(y_{\mathrm {left} })+S(y_{\mathrm {right} })+M(Y_{\mathrm {left} },Y_{\mathrm {right} })}

whereYleft{\displaystyle Y_{\mathrm {left} }} andYright{\displaystyle Y_{\mathrm {right} }} are the sorted versions ofyleft{\displaystyle y_{\mathrm {left} }} andyright{\displaystyle y_{\mathrm {right} }}, andM(,){\displaystyle M(\cdot ,\cdot )} characterizes theBubble Sort swap-equivalent for a merge operation.M(,){\displaystyle M(\cdot ,\cdot )} is computed as depicted in the following pseudo-code:

function M(L[1..n], R[1..m])is    i := 1    j := 1    nSwaps := 0while i ≤ nand j ≤ mdoif R[j] < L[i]then            nSwaps := nSwaps + n − i + 1            j := j + 1else            i := i + 1return nSwaps

A side effect of the above steps is that you end up with both a sorted version ofx{\displaystyle x} and a sorted version ofy{\displaystyle y}. With these, the factorsti{\displaystyle t_{i}} anduj{\displaystyle u_{j}} used to computeτB{\displaystyle \tau _{B}} are easily obtained in a single linear-time pass through the sorted arrays.

Approximating Kendall rank correlation from a stream

[edit]

Efficient algorithms for calculating the Kendall rank correlation coefficient as per the standard estimator haveO(nlogn){\displaystyle O(n\cdot \log {n})} time complexity. However, these algorithms necessitate the availability of all data to determine observation ranks, posing a challenge in sequential data settings where observations are revealed incrementally. Fortunately, algorithms do exist to estimate the Kendall rank correlation coefficient in sequential settings.[19][20] These algorithms haveO(1){\displaystyle O(1)} update time and space complexity, scaling efficiently with the number of observations. Consequently, when processing a batch ofn{\displaystyle n} observations, the time complexity becomesO(n){\displaystyle O(n)}, while space complexity remains a constantO(1){\displaystyle O(1)}.

The first such algorithm[19] presents an approximation to the Kendall rank correlation coefficient based on coarsening the joint distribution of the random variables. Non-stationary data is treated via a moving window approach. This algorithm[19] is simple and is able to handle discrete random variables along with continuous random variables without modification.

The second algorithm[20] is based on Hermite series estimators and utilizes an alternative estimator for the exact Kendall rank correlation coefficient i.e. for the probability of concordance minus the probability of discordance of pairs of bivariate observations. This alternative estimator also serves as an approximation to the standard estimator. This algorithm[20] is only applicable to continuous random variables, but it has demonstrated superior accuracy and potential speed gains compared to the first algorithm described,[19] along with the capability to handle non-stationary data without relying on sliding windows. An efficient implementation of the Hermite series based approach is contained in the R package packagehermiter.[20]

Software Implementations

[edit]

See also

[edit]

References

[edit]
  1. ^abKendall, M. G. (1938). "A New Measure of Rank Correlation".Biometrika.30 (1–2):81–89.doi:10.1093/biomet/30.1-2.81.JSTOR 2332226.
  2. ^Kruskal, W. H. (1958). "Ordinal Measures of Association".Journal of the American Statistical Association.53 (284):814–861.doi:10.2307/2281954.JSTOR 2281954.MR 0100941.
  3. ^Nelsen, R.B. (2001) [1994],"Kendall tau metric",Encyclopedia of Mathematics,EMS Press
  4. ^Prokhorov, A.V. (2001) [1994],"Kendall coefficient of rank correlation",Encyclopedia of Mathematics,EMS Press
  5. ^Valz, Paul D.; McLeod, A. Ian (February 1990)."A Simplified Derivation of the Variance of Kendall's Rank Correlation Coefficient".The American Statistician.44 (1):39–40.doi:10.1080/00031305.1990.10475691.ISSN 0003-1305.
  6. ^Valz, Paul D.; McLeod, A. Ian; Thompson, Mary E. (February 1995)."Cumulant Generating Function and Tail Probability Approximations for Kendall's Score with Tied Rankings".The Annals of Statistics.23 (1):144–160.doi:10.1214/aos/1176324460.ISSN 0090-5364.
  7. ^Hoeffding, Wassily (1992), Kotz, Samuel; Johnson, Norman L. (eds.),"A Class of Statistics with Asymptotically Normal Distribution",Breakthroughs in Statistics: Foundations and Basic Theory, Springer Series in Statistics, New York, NY: Springer, pp. 308–334,doi:10.1007/978-1-4612-0919-5_20,ISBN 978-1-4612-0919-5, retrieved2024-01-19
  8. ^Kendall, M. G. (1949)."Rank and Product-Moment Correlation".Biometrika.36 (1/2):177–193.doi:10.2307/2332540.ISSN 0006-3444.JSTOR 2332540.PMID 18132091.
  9. ^Richard Greiner, (1909),Ueber das Fehlersystem der Kollektiv-maßlehre, Zeitschrift für Mathematik und Physik, Band 57, B. G. Teubner, Leipzig, pages 121-158, 225-260, 337-373.
  10. ^Moran, P. A. P. (1948)."Rank Correlation and Product-Moment Correlation".Biometrika.35 (1/2):203–206.doi:10.2307/2332641.ISSN 0006-3444.JSTOR 2332641.PMID 18867425.
  11. ^Berger, Daniel (2016)."A Proof of Greiner's Equality".SSRN Electronic Journal.doi:10.2139/ssrn.2830471.ISSN 1556-5068.
  12. ^Kendall, M. G. (1945)."The Treatment of Ties in Ranking Problems".Biometrika.33 (3):239–251.doi:10.2307/2332303.PMID 21006841. Retrieved12 November 2024.
  13. ^Alfred Brophy (1986)."An algorithm and program for calculation of Kendall's rank correlation coefficient"(PDF).Behavior Research Methods, Instruments, & Computers.18:45–46.doi:10.3758/BF03200993.S2CID 62601552.
  14. ^IBM (2016).IBM SPSS Statistics 24 Algorithms. IBM. p. 168. Retrieved31 August 2017.
  15. ^abBerry, K. J.; Johnston, J. E.; Zahran, S.; Mielke, P. W. (2009)."Stuart's tau measure of effect size for ordinal variables: Some methodological considerations".Behavior Research Methods.41 (4):1144–1148.doi:10.3758/brm.41.4.1144.PMID 19897822.
  16. ^abcStuart, A. (1953). "The Estimation and Comparison of Strengths of Association in Contingency Tables".Biometrika.40 (1–2):105–110.doi:10.2307/2333101.JSTOR 2333101.
  17. ^Valz, Paul D.; McLeod, A. Ian; Thompson, Mary E. (February 1995)."Cumulant Generating Function and Tail Probability Approximations for Kendall's Score with Tied Rankings".The Annals of Statistics.23 (1):144–160.doi:10.1214/aos/1176324460.ISSN 0090-5364.
  18. ^Knight, W. (1966). "A Computer Method for Calculating Kendall's Tau with Ungrouped Data".Journal of the American Statistical Association.61 (314):436–439.doi:10.2307/2282833.JSTOR 2282833.
  19. ^abcdXiao, W. (2019). "Novel Online Algorithms for Nonparametric Correlations with Application to Analyze Sensor Data".2019 IEEE International Conference on Big Data (Big Data). pp. 404–412.doi:10.1109/BigData47090.2019.9006483.ISBN 978-1-7281-0858-2.S2CID 211298570.
  20. ^abcdeStephanou, M. and Varughese, M (2023). "Hermiter: R package for sequential nonparametric estimation".Computational Statistics.arXiv:2111.14091.doi:10.1007/s00180-023-01382-0.S2CID 244715035.{{cite journal}}: CS1 maint: multiple names: authors list (link)

Further reading

[edit]

External links

[edit]
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Retrieved from "https://en.wikipedia.org/w/index.php?title=Kendall_rank_correlation_coefficient&oldid=1269962713"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp