Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Minimum-distance estimation

From Wikipedia, the free encyclopedia
(Redirected fromMinimum distance estimation)
Method for fitting a statistical model to data

Minimum-distance estimation (MDE) is a conceptual method for fitting a statistical model to data, usually theempirical distribution. Often-used estimators such asordinary least squares can be thought of asspecial cases of minimum-distance estimation.

Whileconsistent andasymptotically normal, minimum-distance estimators are generally notstatistically efficient when compared tomaximum likelihood estimators, because they omit theJacobian usually present in thelikelihood function. This, however, substantially reduces thecomputational complexity of the optimization problem.

Definition

[edit]

LetX1,,Xn{\displaystyle \displaystyle X_{1},\ldots ,X_{n}} be anindependent and identically distributed (iid)randomsample from apopulation withdistributionF(x;θ):θΘ{\displaystyle F(x;\theta )\colon \theta \in \Theta } andΘRk(k1){\displaystyle \Theta \subseteq \mathbb {R} ^{k}(k\geq 1)}.

LetFn(x){\displaystyle \displaystyle F_{n}(x)} be theempirical distribution function based on the sample.

Letθ^{\displaystyle {\hat {\theta }}} be anestimator forθ{\displaystyle \displaystyle \theta }. ThenF(x;θ^){\displaystyle F(x;{\hat {\theta }})} is an estimator forF(x;θ){\displaystyle \displaystyle F(x;\theta )}.

Letd[,]{\displaystyle d[\cdot ,\cdot ]} be afunctional returning some measure of"distance" between its two arguments. The functionald{\displaystyle \displaystyle d} is also called the criterion function.

If there exists aθ^Θ{\displaystyle {\hat {\theta }}\in \Theta } such thatd[F(x;θ^),Fn(x)]=inf{d[F(x;θ),Fn(x)];θΘ}{\displaystyle d[F(x;{\hat {\theta }}),F_{n}(x)]=\inf\{d[F(x;\theta ),F_{n}(x)];\theta \in \Theta \}}, thenθ^{\displaystyle {\hat {\theta }}} is called theminimum-distance estimate ofθ{\displaystyle \displaystyle \theta }.

(Drossos & Philippou 1980, p. 121)

Statistics used in estimation

[edit]

Most theoretical studies of minimum-distance estimation, and most applications, make use of "distance" measures which underlie already-establishedgoodness of fit tests: the test statistic used in one of these tests is used as the distance measure to be minimised. Below are some examples of statistical tests that have been used for minimum-distance estimation.

Chi-square criterion

[edit]

Thechi-square test uses as its criterion the sum, over predefined groups, of the squared difference between the increases of the empirical distribution and the estimated distribution, weighted by the increase in the estimate for that group.

Cramér–von Mises criterion

[edit]

TheCramér–von Mises criterion uses the integral of the squared difference between the empirical and the estimated distribution functions (Parr & Schucany 1980, p. 616).

Kolmogorov–Smirnov criterion

[edit]

TheKolmogorov–Smirnov test uses thesupremum of theabsolute difference between the empirical and the estimated distribution functions (Parr & Schucany 1980, p. 616).

Anderson–Darling criterion

[edit]

TheAnderson–Darling test is similar to the Cramér–von Mises criterion except that the integral is of a weighted version of the squared difference, where the weighting relates the variance of the empirical distribution function (Parr & Schucany 1980, p. 616).

Theoretical results

[edit]

The theory of minimum-distance estimation is related to that for the asymptotic distribution of the corresponding statisticalgoodness of fit tests. Often the cases of theCramér–von Mises criterion, theKolmogorov–Smirnov test and theAnderson–Darling test are treated simultaneously by treating them as special cases of a more general formulation of a distance measure. Examples of the theoretical results that are available are:consistency of the parameter estimates; the asymptotic covariance matrices of the parameter estimates.

See also

[edit]

References

[edit]
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis (see alsoTemplate:Least squares and regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Retrieved from "https://en.wikipedia.org/w/index.php?title=Minimum-distance_estimation&oldid=1230480653"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp