Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Sensitivity index

From Wikipedia, the free encyclopedia
(Redirected fromD prime)
Statistic used in signal detection theory
Figure 1: Bayes-optimal classification error probabilityeb{\displaystyle e_{b}} and Bayes discriminability indexdb{\displaystyle d'_{b}} between two univariate histograms computed from their overlap area. Figure 2: Same computed from the overlap volume of two bivariate histograms. Figure 3: discriminability indices of two univariate normal distributions with unequal variances. The classification boundary is in black. Figure 4: discriminability indices of two bivariate normal distributions with unequal covariance matrices (ellipses are 1 sd error ellipses). Color-bar shows the relative contribution to the discriminability by each dimension. These are computed by numerical methods[1].

Thesensitivity index ordiscriminability index ordetectability index is a dimensionlessstatistic used insignal detection theory. A higher index indicates that the signal can be more readily detected.

Definition

[edit]

The discriminability index is the separation between the means of two distributions (typically the signal and the noise distributions), in units of thestandard deviation.

Equal variances/covariances

[edit]

For twounivariate distributionsa{\displaystyle a} andb{\displaystyle b} with the same standard deviation, it is denoted byd{\displaystyle d'} ('dee-prime'):

d=|μaμb|σ{\displaystyle d'={\frac {\left\vert \mu _{a}-\mu _{b}\right\vert }{\sigma }}}.

In higher dimensions, i.e. with two multivariate distributions with the same variance-covariance matrixΣ{\displaystyle \mathbf {\Sigma } }, (whose symmetric square-root, the standard deviation matrix, isS{\displaystyle \mathbf {S} }), this generalizes to theMahalanobis distance between the two distributions:

d=(μaμb)Σ1(μaμb)=S1(μaμb)=μaμb/σμ{\displaystyle d'={\sqrt {({\boldsymbol {\mu }}_{a}-{\boldsymbol {\mu }}_{b})'\mathbf {\Sigma } ^{-1}({\boldsymbol {\mu }}_{a}-{\boldsymbol {\mu }}_{b})}}=\lVert \mathbf {S} ^{-1}({\boldsymbol {\mu }}_{a}-{\boldsymbol {\mu }}_{b})\rVert =\lVert {\boldsymbol {\mu }}_{a}-{\boldsymbol {\mu }}_{b}\rVert /\sigma _{\boldsymbol {\mu }}},

whereσμ=1/S1μ{\displaystyle \sigma _{\boldsymbol {\mu }}=1/\lVert \mathbf {S} ^{-1}{\boldsymbol {\mu }}\rVert } is the 1d slice of the sd along the unit vectorμ{\displaystyle {\boldsymbol {\mu }}} through the means, i.e. thed{\displaystyle d'} equals thed{\displaystyle d'} along the 1d slice through the means.[1]

For two bivariate distributions with equal variance-covariance, this is given by:

d2=11ρ2(dx2+dy22ρdxdy){\displaystyle {d'}^{2}={\frac {1}{1-\rho ^{2}}}\left({d'}_{x}^{2}+{d'}_{y}^{2}-2\rho {d'}_{x}{d'}_{y}\right)},

whereρ{\displaystyle \rho } is the correlation coefficient, and heredx=μbxμaxσx{\displaystyle d'_{x}={\frac {{\mu _{b}}_{x}-{\mu _{a}}_{x}}{\sigma _{x}}}} anddy=μbyμayσy{\displaystyle d'_{y}={\frac {{\mu _{b}}_{y}-{\mu _{a}}_{y}}{\sigma _{y}}}}, i.e. including the signs of the mean differences instead of the absolute.[1]

d{\displaystyle d'} is also estimated asZ(hit rate)Z(false alarm rate){\displaystyle Z({\text{hit rate}})-Z({\text{false alarm rate}})}.[2]: 8 

Unequal variances/covariances

[edit]

When the two distributions have different standard deviations (or in general dimensions, different covariance matrices), there exist several contending indices, all of which reduce tod{\displaystyle d'} for equal variance/covariance.

Bayes discriminability index

[edit]

This is the maximum (Bayes-optimal) discriminability index for two distributions, based on the amount of their overlap, i.e. the optimal (Bayes) error of classificationeb{\displaystyle e_{b}} by an ideal observer, or its complement, the optimal accuracyab{\displaystyle a_{b}}:

db=2Z(Bayes error rate eb)=2Z(best accuracy rate ab){\displaystyle d'_{b}=-2Z\left({\text{Bayes error rate }}e_{b}\right)=2Z\left({\text{best accuracy rate }}a_{b}\right)},[1]

whereZ{\displaystyle Z} is the inversecumulative distribution function of the standard normal. The Bayes discriminability between univariate or multivariate normal distributions can be numerically computed[1] (Matlab code), and may also be used as an approximation when the distributions are close to normal.

db{\displaystyle d'_{b}} is a positive-definite statistical distance measure that is free of assumptions about the distributions, like theKullback–Leibler divergenceDKL{\displaystyle D_{\text{KL}}}.DKL(a,b){\displaystyle D_{\text{KL}}(a,b)} is asymmetric, whereasdb(a,b){\displaystyle d'_{b}(a,b)} is symmetric for the two distributions. However,db{\displaystyle d'_{b}} does not satisfy thetriangle inequality, so it is not a full metric.[1]

In particular, for a yes/no task between two univariate normal distributions with meansμa,μb{\displaystyle \mu _{a},\mu _{b}} and variancesva>vb{\displaystyle v_{a}>v_{b}}, the Bayes-optimal classification accuracies are:[1]

p(A|a)=p(χ1,vaλ2>vbc),p(B|b)=p(χ1,vbλ2<vac){\displaystyle p(A|a)=p({\chi '}_{1,v_{a}\lambda }^{2}>v_{b}c),\;\;p(B|b)=p({\chi '}_{1,v_{b}\lambda }^{2}<v_{a}c)},

whereχ2{\displaystyle \chi '^{2}} denotes thenon-central chi-squared distribution,λ=(μaμbvavb)2{\displaystyle \lambda =\left({\frac {\mu _{a}-\mu _{b}}{v_{a}-v_{b}}}\right)^{2}}, andc=λ+lnvalnvbvavb{\displaystyle c=\lambda +{\frac {\ln v_{a}-\ln v_{b}}{v_{a}-v_{b}}}}. The Bayes discriminabilitydb=2Z(p(A|a)+p(B|b)2).{\displaystyle d'_{b}=2Z\left({\frac {p\left(A|a\right)+p\left(B|b\right)}{2}}\right).}

db{\displaystyle d'_{b}} can also be computed from theROC curve of a yes/no task between two univariate normal distributions with a single shifting criterion. It can also be computed from the ROC curve of any two distributions (in any number of variables) with a shifting likelihood-ratio, by locating the point on the ROC curve that is farthest from the diagonal.[1]

For a two-interval task between these distributions, the optimal accuracy isab=p(χ~w,k,λ,0,02>0){\displaystyle a_{b}=p\left({\tilde {\chi }}_{{\boldsymbol {w}},{\boldsymbol {k}},{\boldsymbol {\lambda }},0,0}^{2}>0\right)} (χ~2{\displaystyle {\tilde {\chi }}^{2}} denotes thegeneralized chi-squared distribution), wherew=[σs2σn2],k=[11],λ=μsμnσs2σn2[σs2σn2]{\displaystyle {\boldsymbol {w}}={\begin{bmatrix}\sigma _{s}^{2}&-\sigma _{n}^{2}\end{bmatrix}},\;{\boldsymbol {k}}={\begin{bmatrix}1&1\end{bmatrix}},\;{\boldsymbol {\lambda }}={\frac {\mu _{s}-\mu _{n}}{\sigma _{s}^{2}-\sigma _{n}^{2}}}{\begin{bmatrix}\sigma _{s}^{2}&\sigma _{n}^{2}\end{bmatrix}}}.[1] The Bayes discriminabilitydb=2Z(ab){\displaystyle d'_{b}=2Z\left(a_{b}\right)}.

RMS sd discriminability index

[edit]

A common approximate (i.e. sub-optimal) discriminability index that has a closed-form is to take the average of the variances, i.e. the rms of the two standard deviations:da=|μaμb|/σrms{\displaystyle d'_{a}=\left\vert \mu _{a}-\mu _{b}\right\vert /\sigma _{\text{rms}}}[3] (also denoted byda{\displaystyle d_{a}}). It is2{\displaystyle {\sqrt {2}}} times thez{\displaystyle z}-score of the area under thereceiver operating characteristic curve (AUC) of a single-criterion observer. This index is extended to general dimensions as the Mahalanobis distance using the pooled covariance, i.e. withSrms=[(Σa+Σb)/2]12{\displaystyle \mathbf {S} _{\text{rms}}=\left[\left(\mathbf {\Sigma } _{a}+\mathbf {\Sigma } _{b}\right)/2\right]^{\frac {1}{2}}} as the common sd matrix.[1]

Average sd discriminability index

[edit]

Another index isde=|μaμb|/σavg{\displaystyle d'_{e}=\left\vert \mu _{a}-\mu _{b}\right\vert /\sigma _{\text{avg}}}, extended to general dimensions usingSavg=(Sa+Sb)/2{\displaystyle \mathbf {S} _{\text{avg}}=\left(\mathbf {S} _{a}+\mathbf {S} _{b}\right)/2} as the common sd matrix.[1]

Comparison of the indices

[edit]

It has been shown that for two univariate normal distributions,dadedb{\displaystyle d'_{a}\leq d'_{e}\leq d'_{b}}, and for multivariate normal distributions,dade{\displaystyle d'_{a}\leq d'_{e}} still.[1]

Thus,da{\displaystyle d'_{a}} andde{\displaystyle d'_{e}} underestimate the maximum discriminabilitydb{\displaystyle d'_{b}} of univariate normal distributions.da{\displaystyle d'_{a}} can underestimatedb{\displaystyle d'_{b}} by a maximum of approximately 30%. At the limit of high discriminability for univariate normal distributions,de{\displaystyle d'_{e}} converges todb{\displaystyle d'_{b}}. These results often hold true in higher dimensions, but not always.[1] Simpson and Fitter[3] promotedda{\displaystyle d'_{a}} as the best index, particularly for two-interval tasks, but Das and Geisler[1] have shown thatdb{\displaystyle d'_{b}} is the optimal discriminability in all cases, andde{\displaystyle d'_{e}} is often a better closed-form approximation thanda{\displaystyle d'_{a}}, even for two-interval tasks.

The approximate indexdgm{\displaystyle d'_{gm}}, which uses thegeometric mean of the sd's, is less thandb{\displaystyle d'_{b}} at small discriminability, but greater at large discriminability.[1]

Contribution to discriminability by each dimension

[edit]

In general, the contribution to the total discriminability by each dimension or feature may be measured using the amount by which the discriminability drops when that dimension is removed. If the total Bayes discriminability isd{\displaystyle d'} and the Bayes discriminability with dimensioni{\displaystyle i} removed isdi{\displaystyle d'_{-i}}, we can define the contribution of dimensioni{\displaystyle i} asd2di2{\displaystyle {\sqrt {d'^{2}-{d'_{-i}}^{2}}}}. This is the same as the individual discriminability of dimensioni{\displaystyle i} when the covariance matrices are equal and diagonal, but in the other cases, this measure more accurately reflects the contribution of a dimension than its individual discriminability.[1]

Scaling the discriminability of two distributions

[edit]
Scaling the discriminability of two distributions, by linearly interpolating the mean vector and sd matrix (square root of the covariance matrix) of one towards the other. Ellipses are the error ellipses of the two distributions. Black curve is a quadratic boundary that separates the two distributions.[1]

We may sometimes want to scale the discriminability of two data distributions by moving them closer or farther apart. One such case is when we are modeling a detection or classification task, and the model performance exceeds that of the subject or observed data. In that case, we can move the model variable distributions closer together so that it matches the observed performance, while also predicting which specific data points should start overlapping and be misclassified.

There are several ways of doing this. One is to compute the mean vector andcovariance matrix of the two distributions, then effect a linear transformation to interpolate the mean and sd matrix (square root of the covariance matrix) of one of the distributions towards the other.[1]

Another way that is by computing the decision variables of the data points (log likelihood ratio that a point belongs to one distribution vs another) under a multinormal model, then moving these decision variables closer together or farther apart.[1]

See also

[edit]

References

[edit]
  1. ^abcdefghijklmnopqrsDas, Abhranil; Wilson S Geisler (2020). "Methods to integrate multinormals and compute classification measures".arXiv:2012.14331 [stat.ML].
  2. ^MacMillan, N.; Creelman, C. (2005).Detection Theory: A User's Guide. Lawrence Erlbaum Associates.ISBN 9781410611147.
  3. ^abSimpson, A. J.; Fitter, M. J. (1973). "What is the best index of detectability?".Psychological Bulletin.80 (6):481–488.doi:10.1037/h0035203.

External links

[edit]
Retrieved from "https://en.wikipedia.org/w/index.php?title=Sensitivity_index&oldid=1317426696"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp