Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Likelihood-ratio test

From Wikipedia, the free encyclopedia
Statistical test that compares goodness of fit
This article is about the statistical test that compares goodness of fit. For a general description of the likelihood ratio, seeLikelihood ratio. For the use of likelihood ratios in interpreting diagnostic tests, seeLikelihood ratios in diagnostic testing.

Instatistics, thelikelihood-ratio test is ahypothesis test that involves comparing thegoodness of fit of two competingstatistical models, typically one found bymaximization over the entireparameter space and another found after imposing someconstraint, based on the ratio of theirlikelihoods. If the more constrained model (i.e., thenull hypothesis) is supported by theobserved data, the two likelihoods should not differ by more thansampling error.[1] Thus the likelihood-ratio test tests whether this ratio issignificantly different from one, or equivalently whether itsnatural logarithm is significantly different from zero.

The likelihood-ratio test, also known asWilks test,[2] is the oldest of the three classical approaches to hypothesis testing, together with theLagrange multiplier test and theWald test.[3] In fact, the latter two can be conceptualized as approximations to the likelihood-ratio test, and are asymptotically equivalent.[4][5][6] In the case of comparing two models each of which has no unknownparameters, use of the likelihood-ratio test can be justified by theNeyman–Pearson lemma. The lemma demonstrates that the test has the highestpower among all competitors.[7]

Definition

[edit]

General

[edit]

Suppose that we have astatistical model withparameter spaceΘ{\displaystyle \Theta }. Anull hypothesis is often stated by saying that the parameterθ{\displaystyle \theta } lies in a specified subsetΘ0{\displaystyle \Theta _{0}} ofΘ{\displaystyle \Theta }. Thealternative hypothesis is thus thatθ{\displaystyle \theta } lies in thecomplement ofΘ0{\displaystyle \Theta _{0}}, i.e. inΘ  Θ0{\displaystyle \Theta ~\backslash ~\Theta _{0}}, which is denoted byΘ0c{\displaystyle \Theta _{0}^{\text{c}}}. The likelihood ratiotest statistic for the null hypothesisH0:θΘ0{\displaystyle H_{0}\,:\,\theta \in \Theta _{0}} is given by:[8]

λLR=2ln[ supθΘ0L(θ)  supθΘL(θ) ]{\displaystyle \lambda _{\text{LR}}=-2\ln \left[{\frac {~\sup _{\theta \in \Theta _{0}}{\mathcal {L}}(\theta )~}{~\sup _{\theta \in \Theta }{\mathcal {L}}(\theta )~}}\right]}

where the quantity inside the brackets is called the likelihood ratio. Here, thesup{\displaystyle \sup } notation refers to thesupremum. As all likelihoods are positive, and as the constrained maximum cannot exceed the unconstrained maximum, the likelihood ratio isbounded between zero and one.

Often the likelihood-ratio test statistic is expressed as a difference between thelog-likelihoodsλLR=2[(θ0)(θ^)]{\displaystyle \lambda _{\text{LR}}=-2\left[\ell (\theta _{0})-\ell ({\hat {\theta }})\right]}where(θ^)ln[supθΘL(θ)]{\displaystyle \ell ({\hat {\theta }})\equiv \ln \left[\,\sup _{\theta \in \Theta }{\mathcal {L}}(\theta )\,\right]}is the logarithm of the maximized likelihood functionL{\displaystyle {\mathcal {L}}}, and(θ0){\displaystyle \ell (\theta _{0})} is the maximal value in the special case that the null hypothesis is true (but not necessarily a value that maximizesL{\displaystyle {\mathcal {L}}} for the sampled data) andθ0Θ0 and θ^Θ {\displaystyle \theta _{0}\in \Theta _{0}\qquad {\text{ and }}\qquad {\hat {\theta }}\in \Theta ~}denote the respectivearguments of the maxima and the allowed ranges they're embedded in. Multiplying by −2 ensures mathematically that (byWilks' theorem)λLR{\displaystyle \lambda _{\text{LR}}} converges asymptotically to beingχ²-distributed if the null hypothesis happens to be true.[9] Thefinite-sample distributions of likelihood-ratio statistics are generally unknown.[10]

The likelihood-ratio test requires that the models benested – i.e. the more complex model can be transformed into the simpler model by imposing constraints on the former's parameters. Many common test statistics are tests for nested models and can be phrased as log-likelihood ratios or approximations thereof: e.g. theZ-test, theF-test, theG-test, andPearson's chi-squared test; for an illustration with theone-samplet-test, see below.

If the models are not nested, then instead of the likelihood-ratio test, there is a generalization of the test that can usually be used: for details, seerelative likelihood.

Case of simple hypotheses

[edit]
Main article:Neyman–Pearson lemma

A simple-vs.-simple hypothesis test has completely specified models under both the null hypothesis and the alternative hypothesis, which for convenience are written in terms of fixed values of a notional parameterθ{\displaystyle \theta }:

H0:θ=θ0,H1:θ=θ1.{\displaystyle {\begin{aligned}H_{0}&:&\theta =\theta _{0},\\H_{1}&:&\theta =\theta _{1}.\end{aligned}}}In this case, under either hypothesis, the distribution of the data is fully specified: there are no unknown parameters to estimate. For this case, a variant of the likelihood-ratio test is available:[11][12]

Λ(x)= L(θ0x)  L(θ1x) .{\displaystyle \Lambda (x)={\frac {~{\mathcal {L}}(\theta _{0}\mid x)~}{~{\mathcal {L}}(\theta _{1}\mid x)~}}.}

Some older references may use the reciprocal of the function above as the definition.[13] Thus, the likelihood ratio is small if the alternative model is better than the null model.

The likelihood-ratio test provides the decision rule as follows:

The valuesc{\displaystyle c} andq{\displaystyle q} are usually chosen to obtain a specifiedsignificance levelα{\displaystyle \alpha }, via the relationqPr(Λ=cH0) + Pr(Λ<cH0) = α .{\displaystyle q\Pr(\Lambda =c\mid H_{0})~+~\Pr(\Lambda <c\mid H_{0})~=~\alpha ~.}TheNeyman–Pearson lemma states that this likelihood-ratio test is themost powerful among all levelα{\displaystyle \alpha } tests for this case.[7][12]

Interpretation

[edit]

The likelihood ratio is a function of the datax{\displaystyle x}; therefore, it is astatistic, although unusual in that the statistic's value depends on a parameter,θ{\displaystyle \theta }. The likelihood-ratio test rejects the null hypothesis if the value of this statistic is too small. How small is too small depends on the significance level of the test, i.e. on what probability ofType I error is considered tolerable (Type I errors consist of the rejection of a null hypothesis that is true).

Thenumerator corresponds to the likelihood of an observed outcome under thenull hypothesis. Thedenominator corresponds to the maximum likelihood of an observed outcome, varying parameters over the whole parameter space. The numerator of this ratio is less than the denominator; so, the likelihood ratio is between 0 and 1. Low values of the likelihood ratio mean that the observed result was much less likely to occur under the null hypothesis as compared to the alternative. High values of the statistic mean that the observed outcome was nearly as likely to occur under the null hypothesis as the alternative, and so the null hypothesis cannot be rejected.

An example

[edit]

The following example is adapted and abridged fromStuart, Ord & Arnold (1999, §22.2).

Suppose that we have a random sample, of sizen, from a population that is normally-distributed. Both the mean,μ, and the standard deviation,σ, of the population are unknown. We want to test whether the mean is equal to a given value,μ0.

Thus, our null hypothesis isH0μ =μ0  and our alternative hypothesis isH1μμ0 . The likelihood function isL(μ,σx)=(2πσ2)n/2exp(i=1n(xiμ)22σ2).{\displaystyle {\mathcal {L}}(\mu ,\sigma \mid x)=\left(2\pi \sigma ^{2}\right)^{-n/2}\exp \left(-\sum _{i=1}^{n}{\frac {(x_{i}-\mu )^{2}}{2\sigma ^{2}}}\right)\,.}

With some calculation (omitted here), it can then be shown thatλLR=nln(1+t2n1){\displaystyle \lambda _{LR}=n\ln \left(1+{\frac {t^{2}}{n-1}}\right)} wheret is thet-statistic withn − 1 degrees of freedom. Hence we may use the known exact distribution oftn−1 to draw inferences.

Asymptotic distribution: Wilks’ theorem

[edit]
Main article:Wilks' theorem

If the distribution of the likelihood ratio corresponding to a particular null and alternative hypothesis can be explicitly determined then it can directly be used to form decision regions (to sustain or reject the null hypothesis). In most cases, however, the exact distribution of the likelihood ratio corresponding to specific hypotheses is very difficult to determine.[citation needed]

AssumingH0 is true, there is a fundamental result bySamuel S. Wilks: As the sample sizen{\displaystyle n} approaches{\displaystyle \infty }, and if the null hypothesis lies strictly within the interior of the parameter space, the test statisticλLR{\displaystyle \lambda _{\text{LR}}} defined above will beasymptoticallychi-squared distributed (χ2{\displaystyle \chi ^{2}}) withdegrees of freedom equal to the difference in dimensionality ofΘ{\displaystyle \Theta } andΘ0{\displaystyle \Theta _{0}}.[14] This implies that for a great variety of hypotheses, we can calculate the likelihood ratioλ{\displaystyle \lambda } for the data and then compare the observedλLR{\displaystyle \lambda _{\text{LR}}} to theχ2{\displaystyle \chi ^{2}} value corresponding to a desiredstatistical significance as anapproximate statistical test. Other extensions exist.[which?]

See also

[edit]

References

[edit]
  1. ^King, Gary (1989).Unifying Political Methodology : The Likelihood Theory of Statistical Inference. New York: Cambridge University Press. p. 84.ISBN 0-521-36697-6.
  2. ^Li, Bing; Babu, G. Jogesh (2019).A Graduate Course on Statistical Inference. Springer. p. 331.ISBN 978-1-4939-9759-6.
  3. ^Maddala, G. S.; Lahiri, Kajal (2010).Introduction to Econometrics (Fourth ed.). New York: Wiley. p. 200.
  4. ^Buse, A. (1982). "The Likelihood Ratio, Wald, and Lagrange Multiplier Tests: An Expository Note".The American Statistician.36 (3a):153–157.doi:10.1080/00031305.1982.10482817.
  5. ^Pickles, Andrew (1985).An Introduction to Likelihood Analysis. Norwich: W. H. Hutchins & Sons. pp. 24–27.ISBN 0-86094-190-6.
  6. ^Severini, Thomas A. (2000).Likelihood Methods in Statistics. New York: Oxford University Press. pp. 120–121.ISBN 0-19-850650-3.
  7. ^abNeyman, J.;Pearson, E. S. (1933),"On the problem of the most efficient tests of statistical hypotheses"(PDF),Philosophical Transactions of the Royal Society of London A,231 (694–706):289–337,Bibcode:1933RSPTA.231..289N,doi:10.1098/rsta.1933.0009,JSTOR 91247
  8. ^Koch, Karl-Rudolf (1988).Parameter Estimation and Hypothesis Testing in Linear Models. New York: Springer. p. 306.ISBN 0-387-18840-1.
  9. ^Silvey, S.D. (1970).Statistical Inference. London: Chapman & Hall. pp. 112–114.ISBN 0-412-13820-4.
  10. ^Mittelhammer, Ron C.;Judge, George G.; Miller, Douglas J. (2000).Econometric Foundations. New York: Cambridge University Press. p. 66.ISBN 0-521-62394-4.
  11. ^Mood, A.M.; Graybill, F.A.; Boes, D.C. (1974).Introduction to the Theory of Statistics (3rd ed.).McGraw-Hill. §9.2.
  12. ^abStuart, A.; Ord, K.; Arnold, S. (1999),Kendall's Advanced Theory of Statistics, vol. 2A,Arnold, §§20.10–20.13
  13. ^Cox, D. R.;Hinkley, D. V. (1974),Theoretical Statistics,Chapman & Hall, p. 92,ISBN 0-412-12420-3
  14. ^Wilks, S.S. (1938)."The large-sample distribution of the likelihood ratio for testing composite hypotheses".Annals of Mathematical Statistics.9 (1):60–62.doi:10.1214/aoms/1177732360.

Further reading

[edit]

External links

[edit]
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Retrieved from "https://en.wikipedia.org/w/index.php?title=Likelihood-ratio_test&oldid=1316635324"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp