Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Generalized chi-squared distribution

From Wikipedia, the free encyclopedia
Kind of probability distribution
Generalized chi-squared distribution
Probability density function
Generalized chi-square probability density function
Cumulative distribution function
Generalized chi-square cumulative distribution function
Notationχ~(w,k,λ,s,m){\displaystyle {\tilde {\chi }}({\boldsymbol {w}},{\boldsymbol {k}},{\boldsymbol {\lambda }},s,m)}
Parametersw{\displaystyle {\boldsymbol {w}}}, vector of weights of noncentral chi-square components
k{\displaystyle {\boldsymbol {k}}}, vector of degrees of freedom of noncentral chi-square components
λ{\displaystyle {\boldsymbol {\lambda }}}, vector of non-centrality parameters of chi-square components
s{\displaystyle s}, scale of normal term
m{\displaystyle m}, offset
Supportx{[m,+) if wi0,s=0,(,m] if wi0,s=0,R otherwise.{\displaystyle x\in {\begin{cases}[m,+\infty ){\text{ if }}w_{i}\geq 0,s=0,\\(-\infty ,m]{\text{ if }}w_{i}\leq 0,s=0,\\\mathbb {R} {\text{ otherwise.}}\end{cases}}}
PDFno closed-form expression
CDFno closed-form expression
Meanjwj(kj+λj)+m{\displaystyle \sum _{j}w_{j}(k_{j}+\lambda _{j})+m}
Variance2jwj2(kj+2λj)+s2{\displaystyle 2\sum _{j}w_{j}^{2}(k_{j}+2\lambda _{j})+s^{2}}
MGFexp[t(m+jwjλj12wjt)+s2t22]j(12wjt)kj/2{\displaystyle {\frac {\exp \left[t\left(m+\sum _{j}{\frac {w_{j}\lambda _{j}}{1-2w_{j}t}}\right)+{\frac {s^{2}t^{2}}{2}}\right]}{\prod _{j}\left(1-2w_{j}t\right)^{k_{j}/2}}}}
CFexp[it(m+jwjλj12iwjt)s2t22]j(12iwjt)kj/2{\displaystyle {\frac {\exp \left[it\left(m+\sum _{j}{\frac {w_{j}\lambda _{j}}{1-2iw_{j}t}}\right)-{\frac {s^{2}t^{2}}{2}}\right]}{\prod _{j}\left(1-2iw_{j}t\right)^{k_{j}/2}}}}

Inprobability theory andstatistics, thegeneralized chi-squared distribution (orgeneralized chi-square distribution) is the distribution of a quadratic function of amultinormal variable (normal vector), or a linear combination of different normal variables and squares of normal variables. Equivalently, it is also a linear sum of independentnoncentral chi-square variables and anormal variable. There are several other such generalizations for which the same term is sometimes used; some of them are special cases of the family discussed here, for example thegamma distribution.

Definition

[edit]

The generalized chi-squared variable may be described in multiple ways. One is to write it as a weighted sum of independentnoncentral chi-square variablesχ2{\displaystyle {{\chi }'}^{2}} and a standard normal variablez{\displaystyle z}:[1][2][3][4]

χ~(w,k,λ,s,m)=iwiχ2(ki,λi)+sz+m.{\displaystyle {\tilde {\chi }}({\boldsymbol {w}},{\boldsymbol {k}},{\boldsymbol {\lambda }},s,m)=\sum _{i}w_{i}{{\chi }'}^{2}(k_{i},\lambda _{i})+sz+m.}

Here the parameters are the weightswi{\displaystyle w_{i}}, the degrees of freedomki{\displaystyle k_{i}} and non-centralitiesλi{\displaystyle \lambda _{i}} of the constituent non-central chi-squares, and the coefficientss{\displaystyle s} andm{\displaystyle m} of the normal. Some important special cases of this have all weightswi{\displaystyle w_{i}} of the same sign, or have central chi-squared components, or omit the normal term.

Since a non-central chi-squared variable is a sum of squares of normal variables with different means, the generalized chi-square variable is also defined as a sum of squares of independent normal variables, plus an independent normal variable: that is, a quadratic in normal variables.

Another equivalent way is to formulate it as a quadratic form of a normal vectorx{\displaystyle {\boldsymbol {x}}}:[5][6][4][3]

χ~=q(x)=xQ2x+q1x+q0{\displaystyle {\tilde {\chi }}=q({\boldsymbol {x}})={\boldsymbol {x}}'\mathbf {Q_{2}} {\boldsymbol {x}}+{\boldsymbol {q_{1}}}'{\boldsymbol {x}}+q_{0}}.

HereQ2{\displaystyle \mathbf {Q_{2}} } is a matrix,q1{\displaystyle {\boldsymbol {q_{1}}}} is a vector, andq0{\displaystyle q_{0}} is a scalar. These, together with the meanμ{\displaystyle {\boldsymbol {\mu }}} and covariance matrixΣ{\displaystyle \mathbf {\Sigma } } of the normal vectorx{\displaystyle {\boldsymbol {x}}}, parameterize the distribution.

For the most general case, a reduction towards a common standard form can be made by using a representation of the following form:[7]

X=(z+a)TA(z+a)+cTz=(x+b)TD(x+b)+dTx+e,{\displaystyle X=(z+a)^{\mathrm {T} }A(z+a)+c^{\mathrm {T} }z=(x+b)^{\mathrm {T} }D(x+b)+d^{\mathrm {T} }x+e,}

whereD is a diagonal matrix and wherex represents a vector of uncorrelatedstandard normal random variables.

Parameter conversions

[edit]

A generalized chi-square variable or distribution can be parameterized in two ways. The first is in terms of the weightswi{\displaystyle w_{i}}, the degrees of freedomki{\displaystyle k_{i}} and non-centralitiesλi{\displaystyle \lambda _{i}} of the constituent non-central chi-squares, and the coefficientss{\displaystyle s} andm{\displaystyle m} of the added normal term. The second parameterization is using the quadratic form of a normal vector, where the paremeters are the matrixQ2{\displaystyle \mathbf {Q_{2}} }, the vectorq1{\displaystyle {\boldsymbol {q_{1}}}}, and the scalarq0{\displaystyle q_{0}}, and the meanμ{\displaystyle {\boldsymbol {\mu }}} and covariance matrixΣ{\displaystyle \mathbf {\Sigma } } of the normal vector.

The parameters of the first expression (in terms of non-central chi-squares, a normal and a constant) can be calculated in terms of the parameters of the second expression (quadratic form of a normal vector).[6]

The parameters of the second expression (quadratic form of a normal vector) can also be calculated in terms of the parameters of the first expression (in terms of non-central chi-squares, a normal and a constant).[4]

There existsMatlab code to convert from one set of parameters to another.

Support and tails

[edit]

Whens=0{\displaystyle s=0} andwi{\displaystyle w_{i}} are all positive or negative, the quadratic is an ellipse. Then the distribution starts from the pointm{\displaystyle m} at one end, which is called a finite tail. The other end tails off at + or{\displaystyle -\infty } respectively, which is called an infinite tail. Whenwi{\displaystyle w_{i}} have mixed signs, and/or there is a normals{\displaystyle s} term, both tails are infinite, and the support is the entire real line.The methods to compute the CDF and PDF of the distribution behave differently in finite vs. infinite tails (see table below for best method to use in each case).[4][3]

Computing the PDF/CDF/inverse CDF/random numbers

[edit]

The probability density, cumulative distribution, and inverse cumulative distribution functions of a generalized chi-squared variable do not have simple closed-form expressions. But there exist several methods to compute them numerically: Ruben's method,[8] Imhof's method,[9] IFFT method,[4][3] ray method,[4][3] ellipse approximation,[4][3] infinite-tail (or simply 'tail') approximation,[4][3] Pearson and extended Pearson's approximations[4][3] and Liu-Tang-Zhang approximation.

Numerical algorithms[7][2][9][6][4][3] and computer code (Fortran and C,Matlab,R,Python,Julia) have been published that implement some of these methods to compute the PDF, CDF, and inverse CDF, and to generate random numbers.

The following table shows the best methods to use to compute the CDF and PDF for the different parts of the generalized chi-square distribution in different cases.[4][3] 'Tail' refers to the infinite-tail approximation.

χ~{\displaystyle {\tilde {\chi }}} typepartbest cdf/pdf method(s)
ellipse:
wi{\displaystyle w_{i}} same sign,
s=0{\displaystyle s=0}
bodyRuben, Imhof, IFFT, ray
finite tailRuben, ray (ifλi=0{\displaystyle \lambda _{i}=0}), ellipse
infinite tailRuben, ray, tail
not ellipse:
wi{\displaystyle w_{i}} mixed signs,
and/ors0{\displaystyle s\neq 0}
bodyImhof, IFFT, ray
infinite tailsray, tail
sphere:
non-centralχ2{\displaystyle \chi ^{2}}
(only one term)
bodyMatlab'sncx2cdf/ncx2pdf
finite tailncx2cdf/ncx2pdf, ellipse
infinite tailncx2pdf, ray, tail

Asymptotic expressions in the tails

[edit]

Asymptotic expressions for the PDF and CDF of the distribution in the lower or upper infinite tail are given by the infinite-tail approximation:[4][3]

limx±f(x)=a|w|fχk,λ2(xw),{\displaystyle \lim _{x\to \pm \infty }f(x)={\frac {a}{\left|w_{*}\right|}}f_{{\chi '}_{k_{*},\lambda _{*}}^{2}}\!\left({\frac {x}{w_{*}}}\right),}

limxF(x)=limxF¯(x)=a F¯χk,λ2(xw)=a Qk/2(λ,x/w),{\displaystyle \lim _{x\to -\infty }F(x)=\lim _{x\to \infty }{\bar {F}}(x)=a\ {\bar {F}}_{{\chi '}_{k_{*},\lambda _{*}}^{2}}\!\left({\frac {x}{w_{*}}}\right)=a\ Q_{k_{*}/2}({\sqrt {\lambda _{*}}},{\sqrt {x/w_{*}}}),}

where the constant a=em2w+s28w2jexpλjwj2(wwj)(1wjw)kj/2.{\displaystyle {\text{where the constant }}a=e^{{\frac {m}{2w_{*}}}+{\frac {s^{2}}{8w_{*}^{2}}}}\prod _{j\neq *}{\frac {\exp {\frac {\lambda _{j}w_{j}}{2(w_{*}-w_{j})}}}{\left(1-{\frac {w_{j}}{w_{*}}}\right)^{k_{j}/2}}}.}

Here,w{\displaystyle w_{*}} is the largest positive or negative weight if we are looking at the upper or lower tail respectively, andk{\displaystyle k_{*}} andλ{\displaystyle \lambda _{*}} are its corresponding degree and non-centrality.fχk,λ2{\displaystyle f_{{\chi '}_{k_{*},\lambda _{*}}^{2}}} andFχk,λ2{\displaystyle F_{{\chi '}_{k_{*},\lambda _{*}}^{2}}} are the PDF and CDF of the non-central chi-square distribution with parametersk{\displaystyle k_{*}} andλ{\displaystyle \lambda _{*}}.Q{\displaystyle Q} is the Marcum Q-function.

In the far tails, these expressions can be further simplified to ones that are then identical for the pdff(x){\displaystyle f(x)} and tail CDFp(x){\displaystyle p(x)} (which is the CDF at a point in the lower tail, or the complementary CDF at a point in the upper tail):[4][3]

f(x)p(x){(xw)k22ex/2w,if λ=0,(xw)k34ex/2w+λx/w,if λ>0.{\displaystyle f(x)\approx p(x)\approx {\begin{cases}\left({\tfrac {x}{w_{*}}}\right)^{\tfrac {k_{*}-2}{2}}e^{-x/2w_{*}},&{\text{if }}\lambda _{*}=0,\\[1ex]\left({\tfrac {x}{w_{*}}}\right)^{\tfrac {k_{*}-3}{4}}e^{-x/2w_{*}+{\sqrt {\lambda _{*}x/w_{*}}}},&{\text{if }}\lambda _{*}>0.\end{cases}}}

Here again,w{\displaystyle w_{*}} is the largest positive or negative weight if we are looking at the upper or lower tail respectively.

Applications

[edit]

The generalized chi-squared is the distribution ofstatistical estimates in cases where the usualstatistical theory does not hold, as in the examples below.

In model fitting and selection

[edit]

If apredictive model is fitted byleast squares, but theresiduals have eitherautocorrelation orheteroscedasticity, then alternative models can be compared (inmodel selection) by relating changes in thesum of squares to anasymptotically valid generalized chi-squared distribution.[5]

Classifying normal vectors using Gaussian discriminant analysis

[edit]

Ifx{\displaystyle {\boldsymbol {x}}} is a normal vector, its log likelihood is aquadratic form ofx{\displaystyle {\boldsymbol {x}}}, and is hence distributed as a generalized chi-squared. The log likelihood ratio thatx{\displaystyle {\boldsymbol {x}}} arises from one normal distribution versus another is also aquadratic form, so distributed as a generalized chi-squared.[6]

In Gaussian discriminant analysis, samples from multinormal distributions are optimally separated by using aquadratic classifier, a boundary that is a quadratic function (e.g. the curve defined by setting the likelihood ratio between two Gaussians to 1). The classification error rates of different types (false positives and false negatives) are integrals of the normal distributions within the quadratic regions defined by this classifier. Since this is mathematically equivalent to integrating a quadratic form of a normal vector, the result is an integral of a generalized-chi-squared variable.[6]

In signal processing

[edit]

The following application arises in the context ofFourier analysis insignal processing,renewal theory inprobability theory, andmulti-antenna systems inwireless communication. The common factor of these areas is that the sum of exponentially distributed variables is of importance (or identically, the sum of squared magnitudes ofcircularly-symmetric centered complex Gaussian variables).

IfZi{\displaystyle Z_{i}} arekindependent,circularly-symmetric centered complex Gaussian random variables withmean 0 andvarianceσi2{\displaystyle \sigma _{i}^{2}}, then the random variable

Q~=i=1k|Zi|2{\displaystyle {\tilde {Q}}=\sum _{i=1}^{k}|Z_{i}|^{2}}

has a generalized chi-squared distribution of a particular form. The difference from the standard chi-squared distribution is thatZi{\displaystyle Z_{i}} are complex and can have different variances, and the difference from the more general generalized chi-squared distribution is that the relevant scaling matrixA is diagonal. Ifμ=σi2{\displaystyle \mu =\sigma _{i}^{2}} for alli, thenQ~{\displaystyle {\tilde {Q}}}, scaled down byμ/2{\displaystyle \mu /2} (i.e. multiplied by2/μ{\displaystyle 2/\mu }), has achi-squared distribution,χ2(2k){\displaystyle \chi ^{2}(2k)}, also known as anErlang distribution. Ifσi2{\displaystyle \sigma _{i}^{2}} have distinct values for alli, thenQ~{\displaystyle {\tilde {Q}}} has the pdf[10]

f(x;k,σ12,,σk2)=i=1kexσi2σi2j=1,jik(1σj2σi2)for x0.{\displaystyle f(x;k,\sigma _{1}^{2},\ldots ,\sigma _{k}^{2})=\sum _{i=1}^{k}{\frac {e^{-{\frac {x}{\sigma _{i}^{2}}}}}{\sigma _{i}^{2}\prod _{j=1,j\neq i}^{k}\left(1-{\frac {\sigma _{j}^{2}}{\sigma _{i}^{2}}}\right)}}\quad {\text{for }}x\geq 0.}

If there are sets of repeated variances amongσi2{\displaystyle \sigma _{i}^{2}}, assume that they are divided intoM sets, each representing a certain variance value. Denoter=(r1,r2,,rM){\displaystyle \mathbf {r} =(r_{1},r_{2},\dots ,r_{M})} to be the number of repetitions in each group. That is, themth set containsrm{\displaystyle r_{m}} variables that have varianceσm2.{\displaystyle \sigma _{m}^{2}.} It represents an arbitrary linear combination of independentχ2{\displaystyle \chi ^{2}}-distributed random variables with different degrees of freedom:

Q~=m=1Mσm2/2Qm,Qmχ2(2rm).{\displaystyle {\tilde {Q}}=\sum _{m=1}^{M}\sigma _{m}^{2}/2*Q_{m},\quad Q_{m}\sim \chi ^{2}(2r_{m})\,.}

The pdf ofQ~{\displaystyle {\tilde {Q}}} is[11]

f(x;r,σ12,σM2)=m=1M1σm2rmk=1Ml=1rkΨk,l,r(rkl)!(x)rklexσk2, for x0,{\displaystyle f(x;\mathbf {r} ,\sigma _{1}^{2},\dots \sigma _{M}^{2})=\prod _{m=1}^{M}{\frac {1}{\sigma _{m}^{2r_{m}}}}\sum _{k=1}^{M}\sum _{l=1}^{r_{k}}{\frac {\Psi _{k,l,\mathbf {r} }}{(r_{k}-l)!}}(-x)^{r_{k}-l}e^{-{\frac {x}{\sigma _{k}^{2}}}},\quad {\text{ for }}x\geq 0,}

where

Ψk,l,r=(1)rk1iΩk,ljk(ij+rj1ij)(1σj21σk2)(rj+ij),{\displaystyle \Psi _{k,l,\mathbf {r} }=(-1)^{r_{k}-1}\sum _{\mathbf {i} \in \Omega _{k,l}}\prod _{j\neq k}{\binom {i_{j}+r_{j}-1}{i_{j}}}\left({\frac {1}{\sigma _{j}^{2}}}\!-\!{\frac {1}{\sigma _{k}^{2}}}\right)^{-(r_{j}+i_{j})},}

withi=[i1,,iM]T{\displaystyle \mathbf {i} =[i_{1},\ldots ,i_{M}]^{T}} from the setΩk,l{\displaystyle \Omega _{k,l}} ofall partitions ofl1{\displaystyle l-1} (withik=0{\displaystyle i_{k}=0}) defined as

Ωk,l={[i1,,im]Zm;j=1Mij=l1,ik=0,ij0 for all j}.{\displaystyle \Omega _{k,l}=\left\{[i_{1},\ldots ,i_{m}]\in \mathbb {Z} ^{m};\sum _{j=1}^{M}i_{j}\!=l-1,i_{k}=0,i_{j}\geq 0{\text{ for all }}j\right\}.}

See also

[edit]

References

[edit]
  1. ^Davies, R. B. (1973). "Numerical inversion of a characteristic function".Biometrika.60 (2):415–417.doi:10.1093/biomet/60.2.415.
  2. ^abDavies, R. B. (1980). "Algorithm AS155: The distribution of a linear combination ofχ2 random variables".Journal of the Royal Statistical Society. Series C (Applied Statistics).29 (3):323–333.doi:10.2307/2346911.JSTOR 2346911.
  3. ^abcdefghijklDas, Abhranil (2025). "New methods to compute the generalized chi-square distribution".Journal of Statistical Computation and Simulation. Taylor & Francis:1–36.
  4. ^abcdefghijklmDas, Abhranil (2024). "New methods to compute the generalized chi-square distribution".arXiv:2404.05062 [stat.CO].
  5. ^abJones, D. A. (1983). "Statistical analysis of empirical models fitted by optimisation".Biometrika.70 (1):67–88.doi:10.1093/biomet/70.1.67.
  6. ^abcdeDas, Abhranil; Wilson S Geisler (2020). "Methods to integrate multinormals and compute classification measures".arXiv:2012.14331 [stat.ML].
  7. ^abSheil, J.; O'Muircheartaigh, I. (1977). "Algorithm AS106: The distribution of non-negative quadratic forms in normal variables".Journal of the Royal Statistical Society. Series C (Applied Statistics).26 (1):92–98.doi:10.2307/2346884.JSTOR 2346884.
  8. ^Ruben, Harold (1962). "Probability content of regions under spherical normal distributions, IV: The distribution of homogeneous and non-homogeneous quadratic functions of normal variables".The Annals of Mathematical Statistics.33 (2):542–570.doi:10.1214/aoms/1177704580.
  9. ^abImhof, J. P. (1961)."Computing the Distribution of Quadratic Forms in Normal Variables"(PDF).Biometrika.48 (3/4):419–426.doi:10.2307/2332763.JSTOR 2332763.
  10. ^D. Hammarwall, M. Bengtsson, B. Ottersten (2008) "Acquiring Partial CSI for Spatially Selective Transmission by Instantaneous Channel Norm Feedback", IEEE Transactions on Signal Processing, 56, 1188–1204
  11. ^E. Björnson, D. Hammarwall, B. Ottersten (2009)"Exploiting Quantized Channel Norm Feedback through Conditional Statistics in Arbitrarily Correlated MIMO Systems",IEEE Transactions on Signal Processing, 57, 4027–4041

External links

[edit]
Discrete
univariate
with finite
support
with infinite
support
Continuous
univariate
supported on a
bounded interval
supported on a
semi-infinite
interval
supported
on the whole
real line
with support
whose type varies
Mixed
univariate
continuous-
discrete
Multivariate
(joint)
Directional
Degenerate
andsingular
Degenerate
Dirac delta function
Singular
Cantor
Families
Retrieved from "https://en.wikipedia.org/w/index.php?title=Generalized_chi-squared_distribution&oldid=1322460502"
Category:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp