Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Variance-stabilizing transformation

From Wikipedia, the free encyclopedia
Concept in applied statistics

In appliedstatistics, avariance-stabilizing transformation is adata transformation that is specifically chosen either to simplify considerations in graphicalexploratory data analysis or to allow the application of simple regression-based oranalysis of variance techniques.[1]

Overview

[edit]

The aim behind the choice of a variance-stabilizing transformation is to find a simple functionƒ to apply to valuesx in a data set to create new valuesy =ƒ(x) such that the variability of the valuesy is not related to their mean value. For example, suppose that the values x are realizations from differentPoisson distributions: i.e. the distributions each have different mean valuesμ. Then, because for the Poisson distribution the variance is identical to the mean, the variance varies with the mean. However, if the simple variance-stabilizing transformation

y=x{\displaystyle y={\sqrt {x}}\,}

is applied, the sampling variance associated with observation will be nearly constant: seeAnscombe transform for details and some alternative transformations.

While variance-stabilizing transformations are well known for certain parametric families of distributions, such as the Poisson and thebinomial distribution, some types of data analysis proceed more empirically: for example by searching amongpower transformations to find a suitable fixed transformation. Alternatively, if data analysis suggests a functional form for the relation between variance and mean, this can be used to deduce a variance-stabilizing transformation.[2] Thus if, for a meanμ,

var(X)=h(μ),{\displaystyle \operatorname {var} (X)=h(\mu ),\,}

a suitable basis for a variance stabilizing transformation would be

yx1h(μ)dμ,{\displaystyle y\propto \int ^{x}{\frac {1}{\sqrt {h(\mu )}}}\,d\mu ,}

where the arbitrary constant of integration and an arbitrary scaling factor can be chosen for convenience.

Example: relative variance

[edit]

IfX is a positiverandom variable and for some constant, s, the variance is given ash(μ) =s2μ2 then the standard deviation is proportional to the mean, which is called fixedrelative error. In this case, the variance-stabilizing transformation is

y=xdμs2μ2=1sln(x)log(x).{\displaystyle y=\int ^{x}{\frac {d\mu }{\sqrt {s^{2}\mu ^{2}}}}={\frac {1}{s}}\ln(x)\propto \log(x)\,.}

That is, the variance-stabilizing transformation is the logarithmic transformation.

Example: absolute plus relative variance

[edit]

If the variance is given ash(μ) =σ2 +s2μ2 then the variance is dominated by a fixed varianceσ2 when|μ| is small enough and is dominated by the relative variances2μ2 when|μ| is large enough. In this case, the variance-stabilizing transformation is

y=xdμσ2+s2μ2=1sasinhxσ/sasinhxλ.{\displaystyle y=\int ^{x}{\frac {d\mu }{\sqrt {\sigma ^{2}+s^{2}\mu ^{2}}}}={\frac {1}{s}}\operatorname {asinh} {\frac {x}{\sigma /s}}\propto \operatorname {asinh} {\frac {x}{\lambda }}\,.}

That is, the variance-stabilizing transformation is theinverse hyperbolic sine of the scaled valuex /λ forλ =σ /s.

Example: pearson correlation

[edit]

TheFisher transformation is a variance stabilizing transformation for thepearson correlation coefficient.

Relationship to the delta method

[edit]

Here the delta method is presented informally to show the link to variance-stabilizing transformations. For a more formal statement of the delta method, seeDelta method.

LetX{\displaystyle X} be a random variable, withE[X]=μ{\displaystyle E[X]=\mu } andVar(X)=σ2{\displaystyle \operatorname {Var} (X)=\sigma ^{2}}.DefineY=g(X){\displaystyle Y=g(X)}, whereg{\displaystyle g} is a regular function. A first order Taylor approximation forY=g(x){\displaystyle Y=g(x)} is:

Y=g(X)g(μ)+g(μ)(Xμ){\displaystyle Y=g(X)\approx g(\mu )+g'(\mu )(X-\mu )}

From the equation above, we obtain:

E[Y]g(μ){\displaystyle E[Y]\approx g(\mu )} andVar[Y]σ2g(μ)2{\displaystyle \operatorname {Var} [Y]\approx \sigma ^{2}g'(\mu )^{2}}

This approximation method is called delta method.

Consider now a random variableX{\displaystyle X} such thatE[X]=μ{\displaystyle E[X]=\mu } andVar[X]=h(μ){\displaystyle \operatorname {Var} [X]=h(\mu )}.Notice the relation between the variance and the mean, which implies, for example,heteroscedasticity in a linear model. Therefore, the goal is to find a functiong{\displaystyle g} such thatY=g(X){\displaystyle Y=g(X)} has a variance independent (at least approximately) of its expectation.

Imposing the conditionVar[Y]h(μ)g(μ)2=constant{\displaystyle \operatorname {Var} [Y]\approx h(\mu )g'(\mu )^{2}={\text{constant}}}, this equality implies the differential equation:

dgdμ=Ch(μ){\displaystyle {\frac {dg}{d\mu }}={\frac {C}{\sqrt {h(\mu )}}}}

This ordinary differential equation has, byseparation of variables, the following solution:

g(μ)=Cdμh(μ){\displaystyle g(\mu )=\int {\frac {C\,d\mu }{\sqrt {h(\mu )}}}}

This last expression appeared for the first time in aM. S. Bartlett paper.[3]

References

[edit]
  1. ^Everitt, B. S. (2002).The Cambridge Dictionary of Statistics (2nd ed.). CUP.ISBN 0-521-81099-X.
  2. ^Dodge, Y. (2003).The Oxford Dictionary of Statistical Terms. OUP.ISBN 0-19-920613-9.
  3. ^Bartlett, M. S. (1947). "The Use of Transformations".Biometrics.3:39–52.doi:10.2307/3001536.
Retrieved from "https://en.wikipedia.org/w/index.php?title=Variance-stabilizing_transformation&oldid=1328174344"
Category:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp