Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Wald test

From Wikipedia, the free encyclopedia
Statistical test

Instatistics, theWald test (named afterAbraham Wald) assessesconstraints onstatistical parameters based on the weighted distance between theunrestricted estimate and its hypothesized value under thenull hypothesis, where the weight is theprecision of the estimate.[1][2] Intuitively, the larger this weighted distance, the less likely it is that the constraint is true. While thefinite sample distributions of Wald tests are generally unknown,[3]: 138  it has an asymptoticχ2-distribution under the null hypothesis, a fact that can be used to determinestatistical significance.[4]

Together with theLagrange multiplier test and thelikelihood-ratio test, the Wald test is one of three classical approaches tohypothesis testing. An advantage of the Wald test over the other two is that it only requires the estimation of the unrestricted model, which lowers thecomputational burden as compared to the likelihood-ratio test. However, a major disadvantage is that (in finite samples) it is not invariant to changes in the representation of the null hypothesis; in other words, algebraically equivalentexpressions of non-linear parameter restriction can lead to different values of the test statistic.[5][6] That is because the Wald statistic is derived from aTaylor expansion,[7] and different ways of writing equivalent nonlinear expressions lead to nontrivial differences in the corresponding Taylor coefficients.[8] Another aberration, known as the Hauck–Donner effect,[9] can occur inbinomial models when the estimated (unconstrained) parameter is close to theboundary of theparameter space—for instance a fitted probability being extremely close to zero or one—which results in the Wald test no longermonotonically increasing in the distance between the unconstrained and constrained parameter.[10][11]

Mathematical details

[edit]

Under the Wald test, the estimatedθ^{\displaystyle {\hat {\theta }}} that was found as themaximizing argument of the unconstrainedlikelihood function is compared with a hypothesized valueθ0{\displaystyle \theta _{0}}. In particular, the squared differenceθ^θ0{\displaystyle {\hat {\theta }}-\theta _{0}} is weighted by the curvature of the log-likelihood function.

Test on a single parameter

[edit]

If the hypothesis involves only a single parameter restriction, then the Wald statistic takes the following form:

W=(θ^θ0)2var(θ^){\displaystyle W={\frac {{({\widehat {\theta }}-\theta _{0})}^{2}}{\operatorname {var} ({\hat {\theta }})}}}

which under the null hypothesis follows an asymptotic χ2-distribution with one degree of freedom. The square root of the single-restriction Wald statistic can be understood as a (pseudo)t-ratio that is, however, not actuallyt-distributed except for the special case of linear regression withnormally distributed errors.[12] In general, it follows an asymptoticz distribution.[13]

W=θ^θ0se(θ^){\displaystyle {\sqrt {W}}={\frac {{\widehat {\theta }}-\theta _{0}}{\operatorname {se} ({\hat {\theta }})}}}

wherese(θ^){\displaystyle \operatorname {se} ({\widehat {\theta }})} is thestandard error (SE) of themaximum likelihood estimate (MLE), the square root of the variance. There are several ways toconsistently estimate thevariance matrix which in finite samples leads to alternative estimates of standard errors and associated test statistics andp-values.[3]: 129  The validity of still getting an asymptotically normal distribution after plugin-in theMLE estimator ofθ^{\displaystyle {\hat {\theta }}} into theSE relies onSlutsky's theorem.

Test(s) on multiple parameters

[edit]

The Wald test can be used to test a single hypothesis on multiple parameters, as well as to test jointly multiple hypotheses on single/multiple parameters. Letθ^n{\displaystyle {\hat {\theta }}_{n}} be our sample estimator ofP parameters (i.e.,θ^n{\displaystyle {\hat {\theta }}_{n}} is aP×1{\displaystyle P\times 1} vector), which is supposed to follow asymptotically a normal distribution withcovariance matrix V,n(θ^nθ)DN(0,V){\displaystyle {\sqrt {n}}({\hat {\theta }}_{n}-\theta )\,\xrightarrow {\mathcal {D}} \,N(0,V)}.The test ofQ hypotheses on theP parameters is expressed with aQ×P{\displaystyle Q\times P} matrix R:

H0:Rθ=r{\displaystyle H_{0}:R\theta =r}
H1:Rθr{\displaystyle H_{1}:R\theta \neq r}

The distribution of the test statistic under the null hypothesis is

(Rθ^nr)[R(V^n/n)R]1(Rθ^nr)/QDF(Q,nP)nDχQ2/Q,{\displaystyle (R{\hat {\theta }}_{n}-r)'[R({\hat {V}}_{n}/n)R']^{-1}(R{\hat {\theta }}_{n}-r)/Q\quad \xrightarrow {\mathcal {D}} \quad F(Q,n-P)\quad {\xrightarrow[{n\rightarrow \infty }]{\mathcal {D}}}\quad \chi _{Q}^{2}/Q,}

which in turn implies

(Rθ^nr)[R(V^n/n)R]1(Rθ^nr)nDχQ2,{\displaystyle (R{\hat {\theta }}_{n}-r)'[R({\hat {V}}_{n}/n)R']^{-1}(R{\hat {\theta }}_{n}-r)\quad {\xrightarrow[{n\rightarrow \infty }]{\mathcal {D}}}\quad \chi _{Q}^{2},}

whereV^n{\displaystyle {\hat {V}}_{n}} is an estimator of the covariance matrix.[14]

Proof

Supposen(θ^nθ)DN(0,V){\displaystyle {\sqrt {n}}({\hat {\theta }}_{n}-\theta )\,\xrightarrow {\mathcal {D}} \,N(0,V)}. Then, bySlutsky's theorem and by the properties of thenormal distribution, multiplying by R has distribution:

Rn(θ^nθ)=n(Rθ^nr)DN(0,RVR){\displaystyle R{\sqrt {n}}({\hat {\theta }}_{n}-\theta )={\sqrt {n}}(R{\hat {\theta }}_{n}-r)\,\xrightarrow {\mathcal {D}} \,N(0,RVR')}

Recalling that a quadratic form of normal distribution has aChi-squared distribution:

n(Rθ^nr)[RVR]1n(Rθ^nr)DχQ2{\displaystyle {\sqrt {n}}(R{\hat {\theta }}_{n}-r)'[RVR']^{-1}{\sqrt {n}}(R{\hat {\theta }}_{n}-r)\,\xrightarrow {\mathcal {D}} \,\chi _{Q}^{2}}

Rearrangingn finally gives:

(Rθ^nr)[R(V/n)R]1(Rθ^nr)DχQ2{\displaystyle (R{\hat {\theta }}_{n}-r)'[R(V/n)R']^{-1}(R{\hat {\theta }}_{n}-r)\quad \xrightarrow {\mathcal {D}} \quad \chi _{Q}^{2}}

What if the covariance matrix is not known a-priori and needs to be estimated from the data? If we have aconsistent estimatorV^n{\displaystyle {\hat {V}}_{n}} ofV{\displaystyle V} such thatV1V^n{\displaystyle V^{-1}{\hat {V}}_{n}} has a determinant that is distributedχnP2{\displaystyle \chi _{n-P}^{2}}, then by the independence of the covariance estimator and equation above, we have:

(Rθ^nr)[R(V^n/n)R]1(Rθ^nr)/QDF(Q,nP){\displaystyle (R{\hat {\theta }}_{n}-r)'[R({\hat {V}}_{n}/n)R']^{-1}(R{\hat {\theta }}_{n}-r)/Q\quad \xrightarrow {\mathcal {D}} \quad F(Q,n-P)}

Nonlinear hypothesis

[edit]

In the standard form, the Wald test is used to test linear hypotheses that can be represented by a single matrix R. If one wishes to test a non-linear hypothesis of the form:

H0:c(θ)=0{\displaystyle H_{0}:c(\theta )=0}
H1:c(θ)0{\displaystyle H_{1}:c(\theta )\neq 0}

The test statistic becomes:

c(θ^n)[c(θ^n)(V^n/n)c(θ^n)]1c(θ^n)DχQ2{\displaystyle c\left({\hat {\theta }}_{n}\right)'\left[c'\left({\hat {\theta }}_{n}\right)\left({\hat {V}}_{n}/n\right)c'\left({\hat {\theta }}_{n}\right)'\right]^{-1}c\left({\hat {\theta }}_{n}\right)\quad {\xrightarrow {\mathcal {D}}}\quad \chi _{Q}^{2}}

wherec(θ^n){\displaystyle c'({\hat {\theta }}_{n})} is thederivative of c evaluated at the sample estimator. This result is obtained using thedelta method, which uses a first order approximation of the variance.

Non-invariance to re-parameterisations

[edit]

The fact that one uses an approximation of the variance has the drawback that the Wald statistic is not-invariant to a non-linear transformation/reparametrisation of the hypothesis: it can give different answers to the same question, depending on how the question is phrased.[15][5] For example, asking whetherR = 1 is the same as asking whether log R = 0; but the Wald statistic forR = 1 is not the same as the Wald statistic for log R = 0 (because there is in general no neat relationship between the standard errors ofR and log R, so it needs to be approximated).[16]

Alternatives to the Wald test

[edit]

There exist several alternatives to the Wald test, namely thelikelihood-ratio test and theLagrange multiplier test (also known as the score test).Robert F. Engle showed that these three tests, the Wald test, thelikelihood-ratio test and theLagrange multiplier test areasymptotically equivalent.[17] Although they are asymptotically equivalent, in finite samples, they could disagree enough to lead to different conclusions.

There are several reasons to prefer the likelihood ratio test or the Lagrange multiplier to the Wald test:[18][19][20]

  • Non-invariance: As argued above, the Wald test is not invariant under reparametrization, while the likelihood ratio tests will give exactly the same answer whether we work withR, log R or any othermonotonic transformation of R.[5]
  • The other reason is that the Wald test uses two approximations (that we know the standard error orFisher information and the maximum likelihood estimate), whereas the likelihood ratio test depends only on the ratio of likelihood functions under the null hypothesis and alternative hypothesis.
  • The Wald test requires an estimate using the maximizing argument, corresponding to the "full" model. In some cases, the model is simpler under the null hypothesis, so that one might prefer to use thescore test (also called Lagrange multiplier test), which has the advantage that it can be formulated in situations where the variability of the maximizing element is difficult to estimate or computing the estimate according to the maximum likelihood estimator is difficult; e.g. theCochran–Mantel–Haenzel test is a score test.[21]

See also

[edit]

References

[edit]
  1. ^Fahrmeir, Ludwig; Kneib, Thomas; Lang, Stefan; Marx, Brian (2013).Regression : Models, Methods and Applications. Berlin: Springer. p. 663.ISBN 978-3-642-34332-2.
  2. ^Ward, Michael D.; Ahlquist, John S. (2018).Maximum Likelihood for Social Science : Strategies for Analysis.Cambridge University Press. p. 36.ISBN 978-1-316-63682-4.
  3. ^abMartin, Vance; Hurn, Stan; Harris, David (2013).Econometric Modelling with Time Series: Specification, Estimation and Testing. Cambridge University Press.ISBN 978-0-521-13981-6.
  4. ^Davidson, Russell; MacKinnon, James G. (1993). "The Method of Maximum Likelihood : Fundamental Concepts and Notation".Estimation and Inference in Econometrics. New York: Oxford University Press. p. 89.ISBN 0-19-506011-3.
  5. ^abcGregory, Allan W.; Veall, Michael R. (1985)."Formulating Wald Tests of Nonlinear Restrictions".Econometrica.53 (6):1465–1468.doi:10.2307/1913221.JSTOR 1913221.
  6. ^Phillips, P. C. B.; Park, Joon Y. (1988)."On the Formulation of Wald Tests of Nonlinear Restrictions"(PDF).Econometrica.56 (5):1065–1083.doi:10.2307/1911359.JSTOR 1911359.
  7. ^Hayashi, Fumio (2000).Econometrics. Princeton: Princeton University Press. pp. 489–491.ISBN 1-4008-2383-8.,
  8. ^Lafontaine, Francine; White, Kenneth J. (1986). "Obtaining Any Wald Statistic You Want".Economics Letters.21 (1):35–40.doi:10.1016/0165-1765(86)90117-5.
  9. ^Hauck, Walter W. Jr.; Donner, Allan (1977). "Wald's Test as Applied to Hypotheses in Logit Analysis".Journal of the American Statistical Association.72 (360a):851–853.doi:10.1080/01621459.1977.10479969.
  10. ^King, Maxwell L.; Goh, Kim-Leng (2002)."Improvements to the Wald Test".Handbook of Applied Econometrics and Statistical Inference. New York: Marcel Dekker. pp. 251–276.ISBN 0-8247-0652-8.
  11. ^Yee, Thomas William (2022). "On the Hauck–Donner Effect in Wald Tests: Detection, Tipping Points, and Parameter Space Characterization".Journal of the American Statistical Association.117 (540):1763–1774.arXiv:2001.08431.doi:10.1080/01621459.2021.1886936.
  12. ^Cameron, A. Colin; Trivedi, Pravin K. (2005).Microeconometrics : Methods and Applications. New York: Cambridge University Press. p. 137.ISBN 0-521-84805-9.
  13. ^Davidson, Russell; MacKinnon, James G. (1993). "The Method of Maximum Likelihood : Fundamental Concepts and Notation".Estimation and Inference in Econometrics. New York: Oxford University Press. p. 89.ISBN 0-19-506011-3.
  14. ^Harrell, Frank E. Jr. (2001). "Section 9.3.1".Regression modeling strategies. New York: Springer-Verlag.ISBN 0387952322.
  15. ^Fears, Thomas R.; Benichou, Jacques; Gail, Mitchell H. (1996). "A reminder of the fallibility of the Wald statistic".The American Statistician.50 (3):226–227.doi:10.1080/00031305.1996.10474384.
  16. ^Critchley, Frank; Marriott, Paul; Salmon, Mark (1996). "On the Differential Geometry of the Wald Test with Nonlinear Restrictions".Econometrica.64 (5):1213–1222.doi:10.2307/2171963.hdl:1814/524.JSTOR 2171963.
  17. ^Engle, Robert F. (1983). "Wald, Likelihood Ratio, and Lagrange Multiplier Tests in Econometrics". In Intriligator, M. D.; Griliches, Z. (eds.).Handbook of Econometrics. Vol. II. Elsevier. pp. 796–801.ISBN 978-0-444-86185-6.
  18. ^Harrell, Frank E. Jr. (2001). "Section 9.3.3".Regression modeling strategies. New York: Springer-Verlag.ISBN 0387952322.
  19. ^Collett, David (1994).Modelling Survival Data in Medical Research. London: Chapman & Hall.ISBN 0412448807.
  20. ^Pawitan, Yudi (2001).In All Likelihood. New York: Oxford University Press.ISBN 0198507658.
  21. ^Agresti, Alan (2002).Categorical Data Analysis (2nd ed.). Wiley. p. 232.ISBN 0471360937.

Further reading

[edit]

External links

[edit]
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Retrieved from "https://en.wikipedia.org/w/index.php?title=Wald_test&oldid=1215075042"
Category:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp