Movatterモバイル変換


[0]ホーム

URL:


Wikipedia

Skewness

(Redirected fromSkewed distribution)
For the planarity measure in graph theory, seeGraph skewness.

Inprobability theory andstatistics,skewness is a measure of the asymmetry of theprobability distribution of areal-valuedrandom variable about its mean. The skewness value can be positive, zero, negative, or undefined.

Example distribution with positive skewness. These data are from experiments on wheat grass growth.

For aunimodal distribution (a distribution with a single peak), negative skew commonly indicates that thetail is on the left side of the distribution, and positive skew indicates that the tail is on the right. In cases where one tail is long but the other tail is fat, skewness does not obey a simple rule. For example, a zero value in skewness means that the tails on both sides of the mean balance out overall; this is the case for a symmetric distribution but can also be true for an asymmetric distribution where one tail is long and thin, and the other is short but fat. Thus, the judgement on the symmetry of a given distribution by using only its skewness is risky; the distribution shape must be taken into account.

Introduction

edit

Consider the two distributions in the figure just below. Within each graph, the values on the right side of the distribution taper differently from the values on the left side. These tapering sides are calledtails, and they provide a visual means to determine which of the two kinds of skewness a distribution has:

  1. negative skew: The left tail is longer; the mass of the distribution is concentrated on the right of the figure. The distribution is said to beleft-skewed,left-tailed, orskewed to the left, despite the fact that the curve itself appears to be skewed or leaning to the right;left instead refers to the left tail being drawn out and, often, the mean being skewed to the left of a typical center of the data. A left-skewed distribution usually appears as aright-leaning curve.[1]
  2. positive skew: The right tail is longer; the mass of the distribution is concentrated on the left of the figure. The distribution is said to beright-skewed,right-tailed, orskewed to the right, despite the fact that the curve itself appears to be skewed or leaning to the left;right instead refers to the right tail being drawn out and, often, the mean being skewed to the right of a typical center of the data. A right-skewed distribution usually appears as aleft-leaning curve.[1]
 

Skewness in a data series may sometimes be observed not only graphically but by simple inspection of the values. For instance, consider the numeric sequence (49, 50, 51), whose values are evenly distributed around a central value of 50. We can transform this sequence into a negatively skewed distribution by adding a value far below the mean, which is probably a negativeoutlier, e.g. (40, 49, 50, 51). Therefore, the mean of the sequence becomes 47.5, and the median is 49.5. Based on the formula ofnonparametric skew, defined as(μν)/σ,{\displaystyle (\mu -\nu )/\sigma ,}  the skew is negative. Similarly, we can make the sequence positively skewed by adding a value far above the mean, which is probably a positive outlier, e.g. (49, 50, 51, 60), where the mean is 52.5, and the median is 50.5.

As mentioned earlier, a unimodal distribution with zero value of skewness does not imply that this distribution is symmetric necessarily. However, a symmetric unimodal or multimodal distribution always has zero skewness.

 
Example of an asymmetric distribution with zero skewness. This figure serves as a counterexample that zero skewness does not imply symmetric distribution necessarily. (Skewness was calculated by Pearson's moment coefficient of skewness.)

Relationship of mean and median

edit

The skewness is not directly related to the relationship between the mean and median: a distribution with negative skew can have its mean greater than or less than the median, and likewise for positive skew.[2]

 
A general relationship of mean and median under differently skewed unimodal distribution.

In the older notion ofnonparametric skew, defined as(μν)/σ,{\displaystyle (\mu -\nu )/\sigma ,}  whereμ{\displaystyle \mu }  is themean,ν{\displaystyle \nu }  is themedian, andσ{\displaystyle \sigma }  is thestandard deviation, the skewness is defined in terms of this relationship: positive/right nonparametric skew means the mean is greater than (to the right of) the median, while negative/left nonparametric skew means the mean is less than (to the left of) the median. However, the modern definition of skewness and the traditional nonparametric definition do not always have the same sign: while they agree for some families of distributions, they differ in some of the cases, and conflating them is misleading.

If the distribution issymmetric, then the mean is equal to the median, and the distribution has zero skewness.[3] If the distribution is both symmetric andunimodal, then themean =median =mode. This is the case of a coin toss or the series 1,2,3,4,... Note, however, that the converse is not true in general, i.e. zero skewness (defined below) does not imply that the mean is equal to the median.

A 2005 journal article points out:[2]

Many textbooks teach a rule of thumb stating that the mean is right of the median under right skew, and left of the median under left skew. This rule fails with surprising frequency. It can fail inmultimodal distributions, or in distributions where one tail islong but the other isheavy. Most commonly, though, the rule fails in discrete distributions where the areas to the left and right of the median are not equal. Such distributions not only contradict the textbook relationship between mean, median, and skew, they also contradict the textbook interpretation of the median.

 
Distribution of adult residents across US households

For example, in the distribution of adult residents across US households, the skew is to the right. However, since the majority of cases is less than or equal to the mode, which is also the median, the mean sits in the heavier left tail. As a result, the rule of thumb that the mean is right of the median under right skew failed.[2]

Definition

edit

Fisher's moment coefficient of skewness

edit

The skewnessγ1{\displaystyle \gamma _{1}}  of a random variableX is the thirdstandardized momentμ~3{\displaystyle {\tilde {\mu }}_{3}} , defined as:[4][5]

γ1:=μ~3=E[(Xμσ)3]=μ3σ3=E[(Xμ)3](E[(Xμ)2])3/2=κ3κ23/2{\displaystyle \gamma _{1}:={\tilde {\mu }}_{3}=\operatorname {E} \left[\left({\frac {X-\mu }{\sigma }}\right)^{3}\right]={\frac {\mu _{3}}{\sigma ^{3}}}={\frac {\operatorname {E} \left[(X-\mu )^{3}\right]}{\left(\operatorname {E} \left[(X-\mu )^{2}\right]\right)^{3/2}}}={\frac {\kappa _{3}}{\kappa _{2}^{3/2}}}} whereμ is the mean,σ is thestandard deviation, E is theexpectation operator,μ3 is the thirdcentral moment, andκt are thet-thcumulants. It is sometimes referred to asPearson's moment coefficient of skewness,[5] or simply themoment coefficient of skewness,[4] but should not be confused with Pearson's other skewness statistics (see below). The last equality expresses skewness in terms of the ratio of the third cumulantκ3 to the 1.5th power of the second cumulantκ2. This is analogous to the definition ofkurtosis as the fourth cumulant normalized by the square of the second cumulant. The skewness is also sometimes denotedSkew[X].

Ifσ is finite andμ is finite too, then skewness can be expressed in terms of the non-central momentE[X3] by expanding the previous formula:μ~3=E[(Xμσ)3]=E[X3]3μE[X2]+3μ2E[X]μ3σ3=E[X3]3μ(E[X2]μE[X])μ3σ3=E[X3]3μσ2μ3σ3.{\displaystyle {\begin{aligned}{\tilde {\mu }}_{3}&=\operatorname {E} \left[\left({\frac {X-\mu }{\sigma }}\right)^{3}\right]\\&={\frac {\operatorname {E} [X^{3}]-3\mu \operatorname {E} [X^{2}]+3\mu ^{2}\operatorname {E} [X]-\mu ^{3}}{\sigma ^{3}}}\\&={\frac {\operatorname {E} [X^{3}]-3\mu (\operatorname {E} [X^{2}]-\mu \operatorname {E} [X])-\mu ^{3}}{\sigma ^{3}}}\\&={\frac {\operatorname {E} [X^{3}]-3\mu \sigma ^{2}-\mu ^{3}}{\sigma ^{3}}}.\end{aligned}}} 

Examples

edit

Skewness can be infinite, as whenPr[X>x]=x2 for x>1, Pr[X<1]=0{\displaystyle \Pr \left[X>x\right]=x^{-2}{\mbox{ for }}x>1,\ \Pr[X<1]=0} where the third cumulants are infinite, or as whenPr[X<x]={12(1x)3 for x<0,12(1+x)3 for x>0.{\displaystyle \Pr[X<x]={\begin{cases}{\frac {1}{2}}(1-x)^{-3}&{\text{ for }}x<0,\\[2pt]{\frac {1}{2}}(1+x)^{-3}&{\text{ for }}x>0.\end{cases}}} where the third cumulant is undefined.

Examples of distributions with finite skewness include the following.

Sample skewness

edit

For a sample ofn values, two natural estimators of the population skewness are[6]

b1=m3s3=1ni=1n(xix¯)3[1n1i=1n(xix¯)2]3/2{\displaystyle b_{1}={\frac {m_{3}}{s^{3}}}={\frac {{\tfrac {1}{n}}\sum _{i=1}^{n}\left(x_{i}-{\bar {x}}\right)^{3}}{\left[{\tfrac {1}{n-1}}\sum _{i=1}^{n}\left(x_{i}-{\bar {x}}\right)^{2}\right]^{3/2}}}} 

and

g1=m3m23/2=1ni=1n(xix¯)3[1ni=1n(xix¯)2]3/2,{\displaystyle g_{1}={\frac {m_{3}}{m_{2}^{3/2}}}={\frac {{\tfrac {1}{n}}\sum _{i=1}^{n}(x_{i}-{\bar {x}})^{3}}{\left[{\tfrac {1}{n}}\sum _{i=1}^{n}\left(x_{i}-{\bar {x}}\right)^{2}\right]^{3/2}}},} 

wherex¯{\displaystyle {\bar {x}}}  is thesample mean,s is thesample standard deviation,m2 is the (biased) sample second centralmoment, andm3 is the (biased) sample third central moment.[6]g1{\displaystyle g_{1}}  is amethod of moments estimator.

Another common definition of thesample skewness is[6][7]

G1=k3k23/2=n2(n1)(n2)b1=n(n1)n2g1,{\displaystyle {\begin{aligned}G_{1}&={\frac {k_{3}}{k_{2}^{3/2}}}={\frac {n^{2}}{(n-1)(n-2)}}\;b_{1}={\frac {\sqrt {n(n-1)}}{n-2}}\;g_{1},\\\end{aligned}}} 

wherek3{\displaystyle k_{3}}  is the unique symmetric unbiased estimator of the thirdcumulant andk2=s2{\displaystyle k_{2}=s^{2}}  is the symmetric unbiased estimator of the second cumulant (i.e. thesample variance). This adjusted Fisher–Pearson standardized moment coefficientG1{\displaystyle G_{1}}  is the version found inExcel and several statistical packages includingMinitab,SAS andSPSS.[7]

Under the assumption that the underlying random variableX{\displaystyle X}  is normally distributed, it can be shown that all three ratiosb1{\displaystyle b_{1}} ,g1{\displaystyle g_{1}}  andG1{\displaystyle G_{1}}  are unbiased andconsistent estimators of the population skewnessγ1=0{\displaystyle \gamma _{1}=0} , withnb1dN(0,6){\displaystyle {\sqrt {n}}b_{1}\mathrel {\xrightarrow {d} } N(0,6)} , i.e., their distributions converge to a normal distribution with mean 0 and variance 6 (Fisher, 1930).[6] The variance of the sample skewness is thus approximately6/n{\displaystyle 6/n}  for sufficiently large samples. More precisely, in a random sample of sizen from a normal distribution,[8][9]

var(G1)=6n(n1)(n2)(n+1)(n+3).{\displaystyle \operatorname {var} (G_{1})={\frac {6n(n-1)}{(n-2)(n+1)(n+3)}}.} 

In normal samples,b1{\displaystyle b_{1}}  has the smaller variance of the three estimators, with[6]

var(b1)<var(g1)<var(G1).{\displaystyle \operatorname {var} (b_{1})<\operatorname {var} (g_{1})<\operatorname {var} (G_{1}).} 

For non-normal distributions,b1{\displaystyle b_{1}} ,g1{\displaystyle g_{1}}  andG1{\displaystyle G_{1}}  are generallybiased estimators of the population skewnessγ1{\displaystyle \gamma _{1}} ; their expected values can even have the opposite sign from the true skewness. For instance, a mixed distribution consisting of very thin Gaussians centred at −99, 0.5, and 2 with weights 0.01, 0.66, and 0.33 has a skewnessγ1{\displaystyle \gamma _{1}}  of about −9.77, but in a sample of 3G1{\displaystyle G_{1}}  has an expected value of about 0.32, since usually all three samples are in the positive-valued part of the distribution, which is skewed the other way.

Applications

edit

Skewness is a descriptive statistic that can be used in conjunction with thehistogram and the normalquantile plot to characterize the data or distribution.

Skewness indicates the direction and relative magnitude of a distribution's deviation from the normal distribution.

With pronounced skewness, standard statistical inference procedures such as aconfidence interval for a mean will be not only incorrect, in the sense that the true coverage level will differ from the nominal (e.g., 95%) level, but they will also result in unequal error probabilities on each side.

Skewness can be used to obtain approximate probabilities and quantiles of distributions (such asvalue at risk in finance) via theCornish–Fisher expansion.

Many models assume normal distribution; i.e., data are symmetric about the mean. The normal distribution has a skewness of zero. But in reality, data points may not be perfectly symmetric. So, an understanding of the skewness of the dataset indicates whether deviations from the mean are going to be positive or negative.

D'Agostino's K-squared test is agoodness-of-fitnormality test based on sample skewness and sample kurtosis.

Other measures of skewness

edit
 
Comparison ofmean,median andmode of twolog-normal distributions with the same medians and different skewnesses.

Other measures of skewness have been used, including simpler calculations suggested byKarl Pearson[10] (not to be confused with Pearson's moment coefficient of skewness, see above). These other measures are:

Pearson's first skewness coefficient (mode skewness)

edit

The Pearson mode skewness,[11] or first skewness coefficient, is defined as

Pearson's second skewness coefficient (median skewness)

edit

The Pearson median skewness, or second skewness coefficient,[12][13] is defined as

Which is a simple multiple of thenonparametric skew.

Quantile-based measures

edit

Bowley's measure of skewness (from 1901),[14][15] also calledYule's coefficient (from 1912)[16][17] is defined as:Q(3/4)+Q(1/4)2Q(1/2)Q(3/4)Q(1/4)2=Q(3/4)+Q(1/4)2Q(1/2)Q(3/4)Q(1/4),{\displaystyle {\frac {{\frac {Q(3/4)+Q(1/4)}{2}}-Q(1/2)}{\frac {Q(3/4)-Q(1/4)}{2}}}={\frac {Q(3/4)+Q(1/4)-2Q(1/2)}{Q(3/4)-Q(1/4)}},} whereQ is thequantile function (i.e., the inverse of thecumulative distribution function). The numerator is difference between the average of the upper and lower quartiles (a measure of location) and the median (another measure of location), while the denominator is thesemi-interquartile range(Q(3/4)Q(1/4))/2{\displaystyle (Q(3/4)}-{Q(1/4))/2} , which for symmetric distributions is equal to theMAD measure ofdispersion.[citation needed]

Other names for this measure are Galton's measure of skewness,[18] the Yule–Kendall index[19] and the quartile skewness,[20]

Similarly, Kelly's measure of skewness is defined as[21]Q(9/10)+Q(1/10)2Q(1/2)Q(9/10)Q(1/10).{\displaystyle {\frac {Q(9/10)+Q(1/10)-2Q(1/2)}{Q(9/10)-Q(1/10)}}.} 

A more general formulation of a skewness function was described by Groeneveld, R. A. and Meeden, G. (1984):[22][23][24]γ(u)=Q(u)+Q(1u)2Q(1/2)Q(u)Q(1u){\displaystyle \gamma (u)={\frac {Q(u)+Q(1-u)-2Q(1/2)}{Q(u)-Q(1-u)}}} The functionγ(u) satisfies−1 ≤γ(u) ≤ 1 and is well defined without requiring the existence of any moments of the distribution.[22] Bowley's measure of skewness isγ(u) evaluated atu = 3/4 while Kelly's measure of skewness isγ(u) evaluated atu = 9/10. This definition leads to a corresponding overall measure of skewness[23] defined as thesupremum of this over the range1/2 ≤u < 1. Another measure can be obtained by integrating the numerator and denominator of this expression.[22]

Quantile-based skewness measures are at first glance easy to interpret, but they often show significantly larger sample variations than moment-based methods. This means that often samples from a symmetric distribution (like the uniform distribution) have a large quantile-based skewness, just by chance.

Groeneveld and Meeden's coefficient

edit

Groeneveld and Meeden have suggested, as an alternative measure of skewness,[22]

skew(X)=μνE(|Xν|),{\displaystyle \operatorname {skew} (X)={\frac {\mu -\nu }{\operatorname {E} (|X-\nu |)}},} 

whereμ is the mean,ν is the median,|...| is theabsolute value, andE() is the expectation operator. This is closely related in form toPearson's second skewness coefficient.

L-moments

edit

Use ofL-moments in place of moments provides a measure of skewness known as the L-skewness.[25]

Distance skewness

edit

A value of skewness equal to zero does not imply that the probability distribution is symmetric. Thus there is a need for another measure of asymmetry that has this property: such a measure was introduced in 2000.[26] It is calleddistance skewness and denoted bydSkew. IfX is a random variable taking values in thed-dimensional Euclidean space,X has finite expectation,X' is an independent identically distributed copy ofX, and{\displaystyle \|\cdot \|}  denotes the norm in the Euclidean space, then a simplemeasure of asymmetry with respect to location parameterθ isdSkew(X):=1EXXEX+X2θ if Pr(X=θ)1{\displaystyle \operatorname {dSkew} (X):=1-{\frac {\operatorname {E} \|X-X'\|}{\operatorname {E} \|X+X'-2\theta \|}}{\text{ if }}\Pr(X=\theta )\neq 1} anddSkew(X) := 0 forX =θ (with probability 1). Distance skewness is always between 0 and 1, equals 0 if and only ifX is diagonally symmetric with respect toθ (X and2θX have the same probability distribution) and equals 1 if and only ifX is a constantc (cθ{\displaystyle c\neq \theta } ) with probability one.[27] Thus there is a simple consistentstatistical test of diagonal symmetry based on thesample distance skewness:dSkewn(X):=1i,jxixji,jxi+xj2θ.{\displaystyle \operatorname {dSkew} _{n}(X):=1-{\frac {\sum _{i,j}\|x_{i}-x_{j}\|}{\sum _{i,j}\|x_{i}+x_{j}-2\theta \|}}.} 

Medcouple

edit

Themedcouple is a scale-invariant robust measure of skewness, with abreakdown point of 25%.[28] It is themedian of the values of the kernel functionh(xi,xj)=(xixm)(xmxj)xixj{\displaystyle h(x_{i},x_{j})={\frac {(x_{i}-x_{m})-(x_{m}-x_{j})}{x_{i}-x_{j}}}} taken over all couples(xi,xj){\displaystyle (x_{i},x_{j})}  such thatxixmxj{\displaystyle x_{i}\geq x_{m}\geq x_{j}} , wherexm{\displaystyle x_{m}}  is the median of thesample{x1,x2,,xn}{\displaystyle \{x_{1},x_{2},\ldots ,x_{n}\}} . It can be seen as the median of all possible quantile skewness measures.

See also

edit

References

edit

Citations

edit
  1. ^abIllowsky, Barbara; Dean, Susan (27 March 2020)."2.6 Skewness and the Mean, Median, and Mode – Statistics".OpenStax. Retrieved21 December 2022.
  2. ^abcvon Hippel, Paul T. (2005)."Mean, Median, and Skew: Correcting a Textbook Rule".Journal of Statistics Education.13 (2). Archived fromthe original on 20 February 2016.
  3. ^"1.3.5.11. Measures of Skewness and Kurtosis". NIST. Retrieved18 March 2012.
  4. ^ab"Measures of Shape: Skewness and Kurtosis", 2008–2016 by Stan Brown, Oak Road Systems
  5. ^abPearson's moment coefficient of skewness, FXSolver.com
  6. ^abcdeJoanes, D. N.; Gill, C. A. (1998). "Comparing measures of sample skewness and kurtosis".Journal of the Royal Statistical Society, Series D.47 (1):183–189.doi:10.1111/1467-9884.00122.
  7. ^abDoane, David P., and Lori E. Seward."Measuring skewness: a forgotten statistic." Journal of Statistics Education 19.2 (2011): 1-18. (Page 7)
  8. ^Duncan Cramer (1997) Fundamental Statistics for Social Research. Routledge.ISBN 9780415172042 (p 85)
  9. ^Kendall, M.G.; Stuart, A. (1969)The Advanced Theory of Statistics, Volume 1: Distribution Theory, 3rd Edition, Griffin.ISBN 0-85264-141-9 (Ex 12.9)
  10. ^"Archived copy"(PDF). Archived fromthe original(PDF) on 5 July 2010. Retrieved9 April 2010.{{cite web}}: CS1 maint: archived copy as title (link)
  11. ^Weisstein, Eric W."Pearson Mode Skewness".MathWorld.
  12. ^Weisstein, Eric W."Pearson's skewness coefficients".MathWorld.
  13. ^Doane, David P.; Seward, Lori E. (2011)."Measuring Skewness: A Forgotten Statistic?"(PDF).Journal of Statistics Education.19 (2):1–18.doi:10.1080/10691898.2011.11889611.
  14. ^Bowley, A. L. (1901). Elements of Statistics, P.S. King & Son, Laondon. Or in a later edition: BOWLEY, AL. "Elements of Statistics, 4th Edn (New York, Charles Scribner)."(1920).
  15. ^Kenney JF and Keeping ES (1962)Mathematics of Statistics, Pt. 1, 3rd ed., Van Nostrand, (page 102).
  16. ^Yule, George Udny. An introduction to the theory of statistics. C. Griffin, limited, 1912.
  17. ^Groeneveld, Richard A (1991). "An influence function approach to describing the skewness of a distribution".The American Statistician.45 (2):97–102.doi:10.2307/2684367.JSTOR 2684367.
  18. ^Johnson, NL, Kotz, S & Balakrishnan, N (1994) p. 3 and p. 40
  19. ^Wilks DS (1995)Statistical Methods in the Atmospheric Sciences, p 27. Academic Press.ISBN 0-12-751965-3
  20. ^Weisstein, Eric W."Skewness".mathworld.wolfram.com. Retrieved21 November 2019.
  21. ^A.W.L. Pubudu Thilan."Applied Statistics I: Chapter 5: Measures of skewness"(PDF).University of Ruhuna. p. 21.
  22. ^abcdGroeneveld, R.A.; Meeden, G. (1984). "Measuring Skewness and Kurtosis".The Statistician.33 (4):391–399.doi:10.2307/2987742.JSTOR 2987742.
  23. ^abMacGillivray (1992)
  24. ^Hinkley DV (1975) "On power transformations to symmetry",Biometrika, 62, 101–111
  25. ^Hosking, J.R.M. (1992). "Moments or L moments? An example comparing two measures of distributional shape".The American Statistician.46 (3):186–189.doi:10.2307/2685210.JSTOR 2685210.
  26. ^Szekely, G.J. (2000). "Pre-limit and post-limit theorems for statistics", In:Statistics for the 21st Century (eds.C. R. Rao and G. J. Szekely), Dekker, New York, pp. 411–422.
  27. ^Szekely, G. J. and Mori, T. F. (2001) "A characteristic measure of asymmetry and its application for testing diagonal symmetry",Communications in Statistics – Theory and Methods 30/8&9, 1633–1639.
  28. ^G. Brys;M. Hubert; A. Struyf (November 2004). "A Robust Measure of Skewness".Journal of Computational and Graphical Statistics.13 (4):996–1017.doi:10.1198/106186004X12632.S2CID 120919149.

Sources

edit
  • Johnson, NL; Kotz, S; Balakrishnan, N (1994).Continuous Univariate Distributions. Vol. 1 (2 ed.). Wiley.ISBN 0-471-58495-9.
  • MacGillivray, HL (1992). "Shape properties of the g- and h- and Johnson families".Communications in Statistics – Theory and Methods.21 (5):1244–1250.doi:10.1080/03610929208830842.
  • Premaratne, G., Bera, A. K. (2001). Adjusting the Tests for Skewness and Kurtosis for Distributional Misspecifications. Working Paper Number 01-0116, University of Illinois. Forthcoming in Comm in Statistics, Simulation and Computation. 2016 1–15
  • Premaratne, G., Bera, A. K. (2000). Modeling Asymmetry and Excess Kurtosis in Stock Return Data. Office of Research Working Paper Number 00-0123, University of Illinois.
  • Skewness Measures for the Weibull Distribution

External links

edit
Wikiversity has learning resources aboutSkewness
Wikimedia Commons has media related toSkewness (statistics).

[8]ページ先頭

©2009-2025 Movatter.jp