Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Asymptotic theory (statistics)

From Wikipedia, the free encyclopedia
Study of convergence properties of statistical estimators

Instatistics,asymptotic theory, orlarge sample theory, is a framework for assessing properties ofestimators andstatistical tests. Within this framework, it is often assumed that thesample sizen may grow indefinitely; the properties of estimators and tests are then evaluated under the limit ofn → ∞. In practice, a limit evaluation is considered to be approximately valid for large finite sample sizes too.[1]

Overview

[edit]

Most statistical problems begin with a dataset ofsizen. The asymptotic theory proceeds by assuming that it is possible (in principle) to keep collecting additional data, thus that the sample size grows infinitely, i.e.n → ∞. Under the assumption, many results can be obtained that are unavailable for samples of finite size. An example is theweak law of large numbers. The law states that for a sequence ofindependent and identically distributed (IID)random variablesX1,X2, ..., if one value is drawn from each random variable and the average of the firstn values is computed asXn, then theXnconverge in probability to the population meanE[Xi] asn → ∞.[2]

In asymptotic theory, the standard approach isn → ∞. For somestatistical models, slightly different approaches of asymptotics may be used. For example, withpanel data, it is commonly assumed that one dimension in the data remains fixed, whereas the other dimension grows:T = constant andN → ∞, or vice versa.[2]

Besides the standard approach to asymptotics, other alternative approaches exist:

  • Within thelocal asymptotic normality framework, it is assumed that the value of the "true parameter" in the model varies slightly withn, such that then-th model corresponds toθn =θ +h/n. This approach lets us study theregularity of estimators.
  • Whenstatistical tests are studied for their power to distinguish against the alternatives that are close to the null hypothesis, it is done within the so-called "local alternatives" framework: the null hypothesis isH0:θ =θ0 and the alternative isH1:θ =θ0 +h/n. This approach is especially popular for theunit root tests.
  • There are models where the dimension of the parameter spaceΘn slowly expands withn, reflecting the fact that the more observations there are, the more structural effects can be feasibly incorporated in the model.
  • Inkernel density estimation andkernel regression, an additional parameter is assumed—the bandwidthh. In those models, it is typically taken thath → 0 asn → ∞. The rate of convergence must be chosen carefully, though, usuallyhn−1/5.

In many cases, highly accurate results for finite samples can be obtained via numerical methods (i.e. computers); even in such cases, though, asymptotic analysis can be useful. This point was made bySmall (2010, §1.4), as follows.

A primary goal of asymptotic analysis is to obtain a deeperqualitative understanding ofquantitative tools. The conclusions of an asymptotic analysis often supplement the conclusions which can be obtained by numerical methods.

Modes of convergence of random variables

[edit]
Further information:Convergence of random variables

Asymptotic properties

[edit]

Estimators

[edit]

Consistency

[edit]

A sequence of estimates is said to beconsistent, if itconverges in probability to the true value of the parameter being estimated:

θ^n p θ0.{\displaystyle {\hat {\theta }}_{n}\ {\xrightarrow {\overset {}{p}}}\ \theta _{0}.}

That is, roughly speaking with an infinite amount of data theestimator (the formula for generating the estimates) would almost surely give the correct result for the parameter being estimated.[2]

Asymptotic distribution

[edit]

If it is possible to find sequences of non-random constants{an},{bn} (possibly depending on the value ofθ0), and a non-degenerate distributionG such that

bn(θ^nan) d G,{\displaystyle b_{n}({\hat {\theta }}_{n}-a_{n})\ {\xrightarrow {d}}\ G,}

then the sequence of estimatorsθ^n{\displaystyle \textstyle {\hat {\theta }}_{n}} is said to have theasymptotic distributionG.

Most often, the estimators encountered in practice areasymptotically normal, meaning their asymptotic distribution is thenormal distribution, withan =θ0,bn =n, andG =N(0,V):

n(θ^nθ0) d N(0,V).{\displaystyle {\sqrt {n}}({\hat {\theta }}_{n}-\theta _{0})\ {\xrightarrow {d}}\ {\mathcal {N}}(0,V).}

Asymptoticconfidence regions

[edit]

Asymptotic theorems

[edit]

See also

[edit]

References

[edit]
  1. ^Höpfner, R. (2014), Asymptotic Statistics, Walter de Gruyter. 286 pag.ISBN 3110250241,ISBN 978-3110250244
  2. ^abcA. DasGupta (2008),Asymptotic Theory of Statistics and Probability, Springer.ISBN 0387759700,ISBN 978-0387759708

Bibliography

[edit]
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis (see alsoTemplate:Least squares and regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
International
National
Retrieved from "https://en.wikipedia.org/w/index.php?title=Asymptotic_theory_(statistics)&oldid=1073590273"
Category:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp