Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Jackknife resampling

From Wikipedia, the free encyclopedia
Statistical method for resampling
Schematic of jackknife resampling

Instatistics, thejackknife (jackknife cross-validation) is across-validation technique and, therefore, a form ofresampling.It is especially useful forbias andvariance estimation. The jackknife pre-dates other common resampling methods such as thebootstrap. Given a sample of sizen{\displaystyle n}, a jackknifeestimator can be built by aggregating the parameter estimates from each subsample of size(n1){\displaystyle (n-1)} obtained by omitting one observation.[1] The jackknife is a linear approximation of thebootstrap.[2]

The jackknife technique was developed byMaurice Quenouille (1924–1973) from 1949 and refined in 1956.John Tukey expanded on the technique in 1958 and proposed the name "jackknife" because, like a physicaljack-knife (a compact folding knife), it is arough-and-ready tool that can improvise a solution for a variety of problems even though specific problems may be more efficiently solved with a purpose-designed tool.[2]

A simple example: mean estimation

[edit]

The jackknifeestimator of a parameter is found by systematically leaving out each observation from a dataset and calculating the parameter estimate over the remaining observations and then aggregating these calculations.

For example, if the parameter to be estimated is the population mean of random variablex{\displaystyle x}, then for a given set ofi.i.d. observationsx1,...,xn{\displaystyle x_{1},...,x_{n}} the natural estimator is the sample mean:

x¯=1ni=1nxi=1ni[n]xi,{\displaystyle {\bar {x}}={\frac {1}{n}}\sum _{i=1}^{n}x_{i}={\frac {1}{n}}\sum _{i\in [n]}x_{i},}

where the last sum used another way to indicate that the indexi{\displaystyle i} runs over the set[n]={1,,n}{\displaystyle [n]=\{1,\ldots ,n\}}.

Then we proceed as follows: For eachi[n]{\displaystyle i\in [n]} we compute the meanx¯(i){\displaystyle {\bar {x}}_{(i)}} of the jackknife subsample consisting of all but thei{\displaystyle i}-th data point, and this is called thei{\displaystyle i}-th jackknife replicate:

x¯(i)=1n1j[n],jixj,i=1,,n.{\displaystyle {\bar {x}}_{(i)}={\frac {1}{n-1}}\sum _{j\in [n],j\neq i}x_{j},\quad \quad i=1,\dots ,n.}

It could help to think that thesen{\displaystyle n} jackknife replicatesx¯(1),,x¯(n){\displaystyle {\bar {x}}_{(1)},\ldots ,{\bar {x}}_{(n)}} approximate the distribution of the sample meanx¯{\displaystyle {\bar {x}}}. A largern{\displaystyle n} improves the approximation. Then finally to get the jackknife estimator, then{\displaystyle n} jackknife replicates are averaged:

x¯jack=1ni=1nx¯(i).{\displaystyle {\bar {x}}_{\mathrm {jack} }={\frac {1}{n}}\sum _{i=1}^{n}{\bar {x}}_{(i)}.}

One may ask about the bias and the variance ofx¯jack{\displaystyle {\bar {x}}_{\mathrm {jack} }}. From the definition ofx¯jack{\displaystyle {\bar {x}}_{\mathrm {jack} }} as the average of the jackknife replicates one could try to calculate explicitly. The bias is a trivial calculation, but the variance ofx¯jack{\displaystyle {\bar {x}}_{\mathrm {jack} }} is more involved since the jackknife replicates are not independent.

For the special case of the mean, one can show explicitly that the jackknife estimate equals the usual estimate:

1ni=1nx¯(i)=x¯.{\displaystyle {\frac {1}{n}}\sum _{i=1}^{n}{\bar {x}}_{(i)}={\bar {x}}.}

This establishes the identityx¯jack=x¯{\displaystyle {\bar {x}}_{\mathrm {jack} }={\bar {x}}}. Then taking expectations we getE[x¯jack]=E[x¯]=E[x]{\displaystyle E[{\bar {x}}_{\mathrm {jack} }]=E[{\bar {x}}]=E[x]}, sox¯jack{\displaystyle {\bar {x}}_{\mathrm {jack} }} is unbiased, while taking variance we getV[x¯jack]=V[x¯]=V[x]/n{\displaystyle V[{\bar {x}}_{\mathrm {jack} }]=V[{\bar {x}}]=V[x]/n}. However, these properties do not generally hold for parameters other than the mean.

This simple example for the case of mean estimation is just to illustrate the construction of a jackknife estimator, while the real subtleties (and the usefulness) emerge for the case of estimating other parameters, such as higher moments than the mean or other functionals of the distribution.

x¯jack{\displaystyle {\bar {x}}_{\mathrm {jack} }} could be used to construct an empirical estimate of the bias ofx¯{\displaystyle {\bar {x}}}, namelybias^(x¯)jack=c(x¯jackx¯){\displaystyle {\widehat {\operatorname {bias} }}({\bar {x}})_{\mathrm {jack} }=c({\bar {x}}_{\mathrm {jack} }-{\bar {x}})} with some suitable factorc>0{\displaystyle c>0}, although in this case we know thatx¯jack=x¯{\displaystyle {\bar {x}}_{\mathrm {jack} }={\bar {x}}} so this construction does not add any meaningful knowledge, but it gives the correct estimation of the bias (which is zero).

A jackknife estimate of the variance ofx¯{\displaystyle {\bar {x}}} can be calculated from the variance of the jackknife replicatesx¯(i){\displaystyle {\bar {x}}_{(i)}}:[3][4]

var^(x¯)jack=n1ni=1n(x¯(i)x¯jack)2=1n(n1)i=1n(xix¯)2.{\displaystyle {\widehat {\operatorname {var} }}({\bar {x}})_{\mathrm {jack} }={\frac {n-1}{n}}\sum _{i=1}^{n}({\bar {x}}_{(i)}-{\bar {x}}_{\mathrm {jack} })^{2}={\frac {1}{n(n-1)}}\sum _{i=1}^{n}(x_{i}-{\bar {x}})^{2}.}

The left equality defines the estimatorvar^(x¯)jack{\displaystyle {\widehat {\operatorname {var} }}({\bar {x}})_{\mathrm {jack} }} and the right equality is an identity that can be verified directly. Then taking expectations we getE[var^(x¯)jack]=V[x]/n=V[x¯]{\displaystyle E[{\widehat {\operatorname {var} }}({\bar {x}})_{\mathrm {jack} }]=V[x]/n=V[{\bar {x}}]}, so this is an unbiased estimator of the variance ofx¯{\displaystyle {\bar {x}}}.

Estimating the bias of an estimator

[edit]

The jackknife technique can be used to estimate (and correct) the bias of an estimator calculated over the entire sample.

Supposeθ{\displaystyle \theta } is the target parameter of interest, which is assumed to be some functional of the distribution ofx{\displaystyle x}. Based on a finite set of observationsx1,...,xn{\displaystyle x_{1},...,x_{n}}, which is assumed to consist ofi.i.d. copies ofx{\displaystyle x}, the estimatorθ^{\displaystyle {\hat {\theta }}} is constructed:

θ^=fn(x1,,xn).{\displaystyle {\hat {\theta }}=f_{n}(x_{1},\ldots ,x_{n}).}

The value ofθ^{\displaystyle {\hat {\theta }}} is sample-dependent, so this value will change from one random sample to another.

By definition, the bias ofθ^{\displaystyle {\hat {\theta }}} is as follows:

bias(θ^)=E[θ^]θ.{\displaystyle {\text{bias}}({\hat {\theta }})=E[{\hat {\theta }}]-\theta .}

One may wish to compute several values ofθ^{\displaystyle {\hat {\theta }}} from several samples, and average them, to calculate an empirical approximation ofE[θ^]{\displaystyle E[{\hat {\theta }}]}, but this is impossible when there are no "other samples" when the entire set of available observationsx1,...,xn{\displaystyle x_{1},...,x_{n}} was used to calculateθ^{\displaystyle {\hat {\theta }}}. In this kind of situation the jackknife resampling technique may be of help.

We construct the jackknife replicates:

θ^(1)=fn1(x2,x3,xn){\displaystyle {\hat {\theta }}_{(1)}=f_{n-1}(x_{2},x_{3}\ldots ,x_{n})}
θ^(2)=fn1(x1,x3,,xn){\displaystyle {\hat {\theta }}_{(2)}=f_{n-1}(x_{1},x_{3},\ldots ,x_{n})}
{\displaystyle \vdots }
θ^(n)=fn1(x1,x2,,xn1){\displaystyle {\hat {\theta }}_{(n)}=f_{n-1}(x_{1},x_{2},\ldots ,x_{n-1})}

where each replicate is a "leave-one-out" estimate based on the jackknife subsample consisting of all but one of the data points:

θ^(i)=fn1(x1,,xi1,xi+1,,xn)i=1,,n.{\displaystyle {\hat {\theta }}_{(i)}=f_{n-1}(x_{1},\ldots ,x_{i-1},x_{i+1},\ldots ,x_{n})\quad \quad i=1,\dots ,n.}

Then we define their average:

θ^jack=1ni=1nθ^(i){\displaystyle {\hat {\theta }}_{\mathrm {jack} }={\frac {1}{n}}\sum _{i=1}^{n}{\hat {\theta }}_{(i)}}

The jackknife estimate of the bias ofθ^{\displaystyle {\hat {\theta }}} is given by:

bias^(θ^)jack=(n1)(θ^jackθ^){\displaystyle {\widehat {\text{bias}}}({\hat {\theta }})_{\mathrm {jack} }=(n-1)({\hat {\theta }}_{\mathrm {jack} }-{\hat {\theta }})}

and the resulting bias-corrected jackknife estimate ofθ{\displaystyle \theta } is given by:

θ^jack=θ^bias^(θ^)jack=nθ^(n1)θ^jack.{\displaystyle {\hat {\theta }}_{\text{jack}}^{*}={\hat {\theta }}-{\widehat {\text{bias}}}({\hat {\theta }})_{\mathrm {jack} }=n{\hat {\theta }}-(n-1){\hat {\theta }}_{\mathrm {jack} }.}

This removes the bias in the special case that the bias isO(n1){\displaystyle O(n^{-1})} and reduces it toO(n2){\displaystyle O(n^{-2})} in other cases.[2]

Estimating the variance of an estimator

[edit]

The jackknife technique can be also used to estimate the variance of an estimator calculated over the entire sample.

Literature

[edit]

Notes

[edit]
  1. ^Efron 1982, p. 2.
  2. ^abcCameron & Trivedi 2005, p. 375.
  3. ^Efron 1982, p. 14.
  4. ^McIntosh, Avery I."The Jackknife Estimation Method"(PDF).Boston University. Avery I. McIntosh. Archived fromthe original(PDF) on 2016-05-14. Retrieved2016-04-30.: p. 3.

References

[edit]
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis (see alsoTemplate:Least squares and regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Retrieved from "https://en.wikipedia.org/w/index.php?title=Jackknife_resampling&oldid=1318355361"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp