Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Uniformly most powerful test

From Wikipedia, the free encyclopedia
Theoretically optimal hypothesis test

Instatistical hypothesis testing, auniformly most powerful (UMP)test is ahypothesis test which has thegreatestpower1β{\displaystyle 1-\beta } among all possible tests of a givensizeα. For example, according to theNeyman–Pearson lemma, thelikelihood-ratio test is UMP for testing simple (point) hypotheses.

Setting

[edit]

LetX{\displaystyle X} denote a random vector (corresponding to the measurements), taken from aparametrized family ofprobability density functions orprobability mass functionsfθ(x){\displaystyle f_{\theta }(x)}, which depends on the unknown deterministic parameterθΘ{\displaystyle \theta \in \Theta }. The parameter spaceΘ{\displaystyle \Theta } is partitioned into two disjoint setsΘ0{\displaystyle \Theta _{0}} andΘ1{\displaystyle \Theta _{1}}. LetH0{\displaystyle H_{0}} denote the hypothesis thatθΘ0{\displaystyle \theta \in \Theta _{0}}, and letH1{\displaystyle H_{1}} denote the hypothesis thatθΘ1{\displaystyle \theta \in \Theta _{1}}.The binary test of hypotheses is performed using a test functionφ(x){\displaystyle \varphi (x)} with a reject regionR{\displaystyle R} (a subset of measurement space).

φ(x)={1if xR0if xRc{\displaystyle \varphi (x)={\begin{cases}1&{\text{if }}x\in R\\0&{\text{if }}x\in R^{c}\end{cases}}}

meaning thatH1{\displaystyle H_{1}} is in force if the measurementXR{\displaystyle X\in R} and thatH0{\displaystyle H_{0}} is in force if the measurementXRc{\displaystyle X\in R^{c}}.Note thatRRc{\displaystyle R\cup R^{c}} is a disjoint covering of the measurement space.

Formal definition

[edit]

A test functionφ(x){\displaystyle \varphi (x)} is UMP of sizeα{\displaystyle \alpha } if for any other test functionφ(x){\displaystyle \varphi '(x)} satisfying

supθΘ0E[φ(X)|θ]=αα=supθΘ0E[φ(X)|θ]{\displaystyle \sup _{\theta \in \Theta _{0}}\;\operatorname {E} [\varphi '(X)|\theta ]=\alpha '\leq \alpha =\sup _{\theta \in \Theta _{0}}\;\operatorname {E} [\varphi (X)|\theta ]\,}

we have

θΘ1,E[φ(X)|θ]=1β(θ)1β(θ)=E[φ(X)|θ].{\displaystyle \forall \theta \in \Theta _{1},\quad \operatorname {E} [\varphi '(X)|\theta ]=1-\beta '(\theta )\leq 1-\beta (\theta )=\operatorname {E} [\varphi (X)|\theta ].}

The Karlin–Rubin theorem

[edit]

The Karlin–Rubin theorem (named forSamuel Karlin andHerman Rubin) can be regarded as an extension of the Neyman–Pearson lemma for composite hypotheses.[1] Consider a scalar measurement having a probability density function parameterized by a scalar parameterθ, and define the likelihood ratiol(x)=fθ1(x)/fθ0(x){\displaystyle l(x)=f_{\theta _{1}}(x)/f_{\theta _{0}}(x)}.Ifl(x){\displaystyle l(x)} is monotone non-decreasing, inx{\displaystyle x}, for any pairθ1θ0{\displaystyle \theta _{1}\geq \theta _{0}} (meaning that the greaterx{\displaystyle x} is, the more likelyH1{\displaystyle H_{1}} is), then the threshold test:

φ(x)={1if x>x00if x<x0{\displaystyle \varphi (x)={\begin{cases}1&{\text{if }}x>x_{0}\\0&{\text{if }}x<x_{0}\end{cases}}}
wherex0{\displaystyle x_{0}} is chosen such thatEθ0φ(X)=α{\displaystyle \operatorname {E} _{\theta _{0}}\varphi (X)=\alpha }

is the UMP test of sizeα for testingH0:θθ0 vs. H1:θ>θ0.{\displaystyle H_{0}:\theta \leq \theta _{0}{\text{ vs. }}H_{1}:\theta >\theta _{0}.}

Note that exactly the same test is also UMP for testingH0:θ=θ0 vs. H1:θ>θ0.{\displaystyle H_{0}:\theta =\theta _{0}{\text{ vs. }}H_{1}:\theta >\theta _{0}.}

Important case: exponential family

[edit]

Although the Karlin-Rubin theorem may seem weak because of its restriction to scalar parameter and scalar measurement, it turns out that there exist a host of problems for which the theorem holds. In particular, the one-dimensionalexponential family ofprobability density functions orprobability mass functions with

fθ(x)=g(θ)h(x)exp(η(θ)T(x)){\displaystyle f_{\theta }(x)=g(\theta )h(x)\exp(\eta (\theta )T(x))}

has a monotone non-decreasing likelihood ratio in thesufficient statisticT(x){\displaystyle T(x)}, provided thatη(θ){\displaystyle \eta (\theta )} is non-decreasing.

Example

[edit]

LetX=(X0,,XM1){\displaystyle X=(X_{0},\ldots ,X_{M-1})} denote i.i.d. normally distributedN{\displaystyle N}-dimensional random vectors with meanθm{\displaystyle \theta m} and covariance matrixR{\displaystyle R}. We then have

fθ(X)=(2π)MN/2|R|M/2exp{12n=0M1(Xnθm)TR1(Xnθm)}=(2π)MN/2|R|M/2exp{12n=0M1(θ2mTR1m)}exp{12n=0M1XnTR1Xn}exp{θmTR1n=0M1Xn}{\displaystyle {\begin{aligned}f_{\theta }(X)={}&(2\pi )^{-MN/2}|R|^{-M/2}\exp \left\{-{\frac {1}{2}}\sum _{n=0}^{M-1}(X_{n}-\theta m)^{T}R^{-1}(X_{n}-\theta m)\right\}\\[4pt]={}&(2\pi )^{-MN/2}|R|^{-M/2}\exp \left\{-{\frac {1}{2}}\sum _{n=0}^{M-1}\left(\theta ^{2}m^{T}R^{-1}m\right)\right\}\\[4pt]&\exp \left\{-{\frac {1}{2}}\sum _{n=0}^{M-1}X_{n}^{T}R^{-1}X_{n}\right\}\exp \left\{\theta m^{T}R^{-1}\sum _{n=0}^{M-1}X_{n}\right\}\end{aligned}}}

which is exactly in the form of the exponential family shown in the previous section, with the sufficient statistic being

T(X)=mTR1n=0M1Xn.{\displaystyle T(X)=m^{T}R^{-1}\sum _{n=0}^{M-1}X_{n}.}

Thus, we conclude that the test

φ(T)={1T>t00T<t0Eθ0φ(T)=α{\displaystyle \varphi (T)={\begin{cases}1&T>t_{0}\\0&T<t_{0}\end{cases}}\qquad \operatorname {E} _{\theta _{0}}\varphi (T)=\alpha }

is the UMP test of sizeα{\displaystyle \alpha } for testingH0:θθ0{\displaystyle H_{0}:\theta \leqslant \theta _{0}} vs.H1:θ>θ0{\displaystyle H_{1}:\theta >\theta _{0}}

Further discussion

[edit]

In general, UMP tests do not exist for vector parameters or for two-sided tests (a test in which one hypothesis lies on both sides of the alternative). The reason is that in these situations, the most powerful test of a given size for one possible value of the parameter (e.g. forθ1{\displaystyle \theta _{1}} whereθ1>θ0{\displaystyle \theta _{1}>\theta _{0}}) is different from the most powerful test of the same size for a different value of the parameter (e.g. forθ2{\displaystyle \theta _{2}} whereθ2<θ0{\displaystyle \theta _{2}<\theta _{0}}). As a result, no test isuniformly most powerful in these situations.

This article includes a list ofgeneral references, butit lacks sufficient correspondinginline citations. Please help toimprove this article byintroducing more precise citations.(November 2010) (Learn how and when to remove this message)

References

[edit]
  1. ^Casella, G.; Berger, R.L. (2008),Statistical Inference, Brooks/Cole.ISBN 0-495-39187-5 (Theorem 8.3.17)

Further reading

[edit]
  • Ferguson, T. S. (1967). "Sec. 5.2:Uniformly most powerful tests".Mathematical Statistics: A decision theoretic approach. New York: Academic Press.
  • Mood, A. M.; Graybill, F. A.; Boes, D. C. (1974). "Sec. IX.3.2:Uniformly most powerful tests".Introduction to the theory of statistics (3rd ed.). New York: McGraw-Hill.
  • L. L. Scharf,Statistical Signal Processing, Addison-Wesley, 1991, section 4.7.
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis (see alsoTemplate:Least squares and regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Retrieved from "https://en.wikipedia.org/w/index.php?title=Uniformly_most_powerful_test&oldid=1312437506"
Category:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp