Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Hermite distribution

From Wikipedia, the free encyclopedia
Statistical probability Distribution for discrete event counts
Hermite
Probability mass function
PMF Hermite
The horizontal axis is the indexk, the number of occurrences. The function is only defined at integer values ofk. The connecting lines are only guides for the eye.
Cumulative distribution function
Plot of the Hermite CDF
The horizontal axis is the indexk, the number of occurrences. The CDF is discontinuous at the integers ofk and flat everywhere else because a variable that is Hermite distributed only takes on integer values.
NotationHerm(a1,a2){\displaystyle \operatorname {Herm} (a_{1},a_{2})\,}
Parametersa1 ≥ 0,a2 ≥ 0
Supportx ∈ { 0, 1, 2, ... }
PMFxe(a1+a2)j=0x/2a1x2ja2j(x2j)!j!{\displaystyle x\mapsto e^{-(a_{1}+a_{2})}\sum _{j=0}^{\lfloor x/2\rfloor }{\frac {a_{1}^{x-2j}a_{2}^{j}}{(x-2j)!j!}}}
CDFxea1+a2i=0xj=0i/2a1i2ja2j(i2j)!j!{\displaystyle x\mapsto e^{-a_{1}+a_{2}}\sum _{i=0}^{\lfloor x\rfloor }\sum _{j=0}^{\lfloor i/2\rfloor }{\frac {a_{1}^{i-2j}a_{2}^{j}}{(i-2j)!j!}}}
Meana1+2a2{\displaystyle a_{1}+2a_{2}}
Variancea1+4a2{\displaystyle a_{1}+4a_{2}}
Skewnessa1+8a2(a1+4a2)3/2{\displaystyle {\frac {a_{1}+8a_{2}}{(a_{1}+4a_{2})^{3/2}}}}
Excess kurtosisa1+16a2(a1+4a2)2{\displaystyle {\frac {a_{1}+16a_{2}}{(a_{1}+4a_{2})^{2}}}}
MGFexp(a1(et1)+a2(e2t1)){\displaystyle \exp(a_{1}(e^{t}-1)+a_{2}(e^{2t}-1))\,}
CFexp(a1(eti1)+a2(e2ti1)){\displaystyle \exp(a_{1}(e^{ti}-1)+a_{2}(e^{2ti}-1))\,}
PGFexp(a1(s1)+a2(s21)){\displaystyle \exp(a_{1}(s-1)+a_{2}(s^{2}-1))\,}

Inprobability theory andstatistics, theHermite distribution, named afterCharles Hermite, is adiscrete probability distribution used to modelcount data with more than one parameter. This distribution is flexible in terms of its ability to allow a moderateover-dispersion in the data.

The authors C. D. Kemp andA. W. Kemp[1] have called it "Hermite distribution" from the fact itsprobability function and themoment generating function can be expressed in terms of the coefficients of (modified)Hermite polynomials.

History

[edit]

The distribution first appeared in the paperApplications of Mathematics to Medical Problems,[2] byAnderson Gray McKendrick in 1926. In this work the author explains several mathematical methods that can be applied to medical research. In one of this methods he considered thebivariate Poisson distribution and showed that the distribution of the sum of two correlated Poisson variables follow a distribution that later would be known as Hermite distribution.

As a practical application, McKendrick considered the distribution of counts ofbacteria inleucocytes. Using themethod of moments he fitted the data with the Hermite distribution and found the model more satisfactory than fitting it with aPoisson distribution.

The distribution was formally introduced and published by C. D. Kemp and Adrienne W. Kemp in 1965 in their workSome Properties of ‘Hermite’ Distribution. The work is focused on the properties of this distribution for instance a necessary condition on the parameters and theirmaximum likelihood estimators (MLE), the analysis of theprobability generating function (PGF) and how it can be expressed in terms of the coefficients of (modified)Hermite polynomials. An example they have used in this publication is the distribution of counts of bacteria in leucocytes that used McKendrick but Kemp and Kemp estimate the model using themaximum likelihood method.

Hermite distribution is a special case of discretecompound Poisson distribution with only two parameters.[3][4]

The same authors published in 1966 the paperAn alternative Derivation of the Hermite Distribution.[5] In this work established that the Hermite distribution can be obtained formally bycompounding aPoisson distribution with anormal distribution.

In 1971, Y. C. Patel[6] did a comparative study of various estimation procedures for the Hermite distribution in his doctoral thesis. It included maximum likelihood, moment estimators, mean and zero frequency estimators and the method of even points.

In 1974, Gupta and Jain[7] did a research on a generalized form of Hermite distribution.

Definition

[edit]

Probability mass function

[edit]

LetX1 andX2 be two independent Poisson variables with parametersa1 anda2. Theprobability distribution of therandom variableY =X1 + 2X2 is the Hermite distribution with parametersa1 anda2 andprobability mass function is given by[8]

pn=P(Y=n)=e(a1+a2)j=0n/2a1n2ja2j(n2j)!j!{\displaystyle p_{n}=P(Y=n)=e^{-(a_{1}+a_{2})}\sum _{j=0}^{\lfloor n/2\rfloor }{\frac {a_{1}^{n-2j}a_{2}^{j}}{(n-2j)!j!}}}

where

Theprobability generating function of the probability mass is,[8]

GY(s)=n=0pnsn=exp(a1(s1)+a2(s21)){\displaystyle G_{Y}(s)=\sum _{n=0}^{\infty }p_{n}s^{n}=\exp(a_{1}(s-1)+a_{2}(s^{2}-1))}

Notation

[edit]

When arandom variableY =X1 + 2X2 is distributed by an Hermite distribution, whereX1 andX2 are two independent Poisson variables with parametersa1 anda2, we write

Y Herm(a1,a2){\displaystyle Y\ \sim \operatorname {Herm} (a_{1},a_{2})\,}

Properties

[edit]

Moment and cumulant generating functions

[edit]

Themoment generating function of a random variableX is defined as the expected value ofet, as a function of the real parametert. For an Hermite distribution with parametersX1 andX2, the moment generating function exists and is equal to

M(t)=G(et)=exp(a1(et1)+a2(e2t1)){\displaystyle M(t)=G(e^{t})=\exp(a_{1}(e^{t}-1)+a_{2}(e^{2t}-1))}

Thecumulant generating function is the logarithm of the moment generating function and is equal to[4]

K(t)=log(M(t))=a1(et1)+a2(e2t1){\displaystyle K(t)=\log(M(t))=a_{1}(e^{t}-1)+a_{2}(e^{2t}-1)}

If we consider the coefficient of (it)rr! in the expansion ofK(t) we obtain ther-cumulant

kn=a1+2na2{\displaystyle k_{n}=a_{1}+2^{n}a_{2}}

Hence themean and the succeeding threemoments about it are

OrderMomentCumulant
1μ1=k1=a1+2a2{\displaystyle \mu _{1}=k_{1}=a_{1}+2a_{2}}μ{\displaystyle \mu }
2μ2=k2=a1+4a2{\displaystyle \mu _{2}=k_{2}=a_{1}+4a_{2}}σ2{\displaystyle \sigma ^{2}}
3μ3=k3=a1+8a2{\displaystyle \mu _{3}=k_{3}=a_{1}+8a_{2}}k3{\displaystyle k_{3}}
4μ4=k4+3k22=a1+16a2+3(a1+4a2)2{\displaystyle \mu _{4}=k_{4}+3k_{2}^{2}=a_{1}+16a_{2}+3(a_{1}+4a_{2})^{2}}k4{\displaystyle k_{4}}

Skewness

[edit]

Theskewness is the third moment centered around the mean divided by the 3/2 power of thestandard deviation, and for the hermite distribution is,[4]

γ1=μ3μ23/2=a1+8a2(a1+4a2)3/2{\displaystyle \gamma _{1}={\frac {\mu _{3}}{\mu _{2}^{3/2}}}={\frac {a_{1}+8a_{2}}{(a_{1}+4a_{2})^{3/2}}}}

Kurtosis

[edit]

Thekurtosis is the fourth moment centered around the mean, divided by the square of thevariance, and for the Hermite distribution is,[4]

β2=μ4μ22=a1+16a2+3(a1+4a2)2(a1+4a2)2=a1+16a2(a1+4a2)2+3{\displaystyle \beta _{2}={\frac {\mu _{4}}{\mu _{2}^{2}}}={\frac {a_{1}+16a_{2}+3(a_{1}+4a_{2})^{2}}{(a_{1}+4a_{2})^{2}}}={\frac {a_{1}+16a_{2}}{(a_{1}+4a_{2})^{2}}}+3}

Theexcess kurtosis is just a correction to make the kurtosis of the normal distribution equal to zero, and it is the following,

γ2=μ4μ223=a1+16a2(a1+4a2)2{\displaystyle \gamma _{2}={\frac {\mu _{4}}{\mu _{2}^{2}}}-3={\frac {a_{1}+16a_{2}}{(a_{1}+4a_{2})^{2}}}}

Characteristic function

[edit]

In a discrete distribution thecharacteristic function of any real-valued random variable is defined as theexpected value ofeitX{\displaystyle e^{itX}}, wherei is the imaginary unit andt ∈ R

ϕ(t)=E[eitX]=j=0eijtP[X=j]{\displaystyle \phi (t)=E[e^{itX}]=\sum _{j=0}^{\infty }e^{ijt}P[X=j]}

This function is related to the moment-generating function viaϕx(t)=MX(it){\displaystyle \phi _{x}(t)=M_{X}(it)}. Hence for this distribution the characteristic function is,[1]

ϕx(t)=exp(a1(eit1)+a2(e2it1)){\displaystyle \phi _{x}(t)=\exp(a_{1}(e^{it}-1)+a_{2}(e^{2it}-1))}

Cumulative distribution function

[edit]

Thecumulative distribution function is,[1]

F(x;a1,a2)=P(Xx)=exp((a1+a2))i=0xj=0[i/2]a1i2ja2j(i2j)!j!{\displaystyle {\begin{aligned}F(x;a_{1},a_{2})&=P(X\leq x)\\&=\exp(-(a_{1}+a_{2}))\sum _{i=0}^{\lfloor x\rfloor }\sum _{j=0}^{[i/2]}{\frac {a_{1}^{i-2j}a_{2}^{j}}{(i-2j)!j!}}\end{aligned}}}

Other properties

[edit]
Example of a multi-modal data, Hermite distribution(0.1,1.5).
d=Var(Y)E(Y)=a1+4a2a1+2a2=1+2a2a1+2a2{\displaystyle d={\frac {\operatorname {Var} (Y)}{\operatorname {E} (Y)}}={\frac {a_{1}+4a_{2}}{a_{1}+2a_{2}}}=1+{\frac {2a_{2}}{a_{1}+2a_{2}}}}

Parameter estimation

[edit]

Method of moments

[edit]

Themean and thevariance of the Hermite distribution areμ=a1+2a2{\displaystyle \mu =a_{1}+2a_{2}} andσ2=a1+4a2{\displaystyle \sigma ^{2}=a_{1}+4a_{2}}, respectively. So we have these two equation,

{x¯=a1+2a2σ2=a1+4a2{\displaystyle {\begin{cases}{\bar {x}}=a_{1}+2a_{2}\\\sigma ^{2}=a_{1}+4a_{2}\end{cases}}}

Solving these two equation we get the moment estimatorsa1^{\displaystyle {\hat {a_{1}}}} anda2^{\displaystyle {\hat {a_{2}}}} ofa1 anda2.[6]

a1^=2x¯σ2{\displaystyle {\hat {a_{1}}}=2{\bar {x}}-\sigma ^{2}}
a2^=σ2x^2{\displaystyle {\hat {a_{2}}}={\frac {\sigma ^{2}-{\hat {x}}}{2}}}

Sincea1 anda2 both are positive, the estimatora1^{\displaystyle {\hat {a_{1}}}} anda2^{\displaystyle {\hat {a_{2}}}} are admissible (≥ 0) only if,x¯<σ2<2x¯{\displaystyle {\bar {x}}<\sigma ^{2}<2{\bar {x}}}.

Maximum likelihood

[edit]

Given a sampleX1, ...,Xm areindependent random variables each having an Hermite distribution we wish to estimate the value of the parametersa1^{\displaystyle {\hat {a_{1}}}} anda2^{\displaystyle {\hat {a_{2}}}}. We know that the mean and the variance of the distribution areμ=a1+2a2{\displaystyle \mu =a_{1}+2a_{2}} andσ2=a1+4a2{\displaystyle \sigma ^{2}=a_{1}+4a_{2}}, respectively. Using these two equation,

{a1=μ(2d)a2=μ(d1)2{\displaystyle {\begin{cases}a_{1}=\mu (2-d)\\[4pt]a_{2}={\dfrac {\mu (d-1)}{2}}\end{cases}}}

We can parameterize the probability function by μ andd

P(X=x)=exp((μ(2d)+μ(d1)2))j=0[x/2](μ(2d))x2j(μ(d1)2)j(x2j)!j!{\displaystyle P(X=x)=\exp \left(-\left(\mu (2-d)+{\frac {\mu (d-1)}{2}}\right)\right)\sum _{j=0}^{[x/2]}{\frac {(\mu (2-d))^{x-2j}\left({\frac {\mu (d-1)}{2}}\right)^{j}}{(x-2j)!j!}}}

Hence thelog-likelihood function is,[9]

L(x1,,xm;μ,d)=log(L(x1,,xm;μ,d))=mμ(1+d12)+log(μ(2d))i=1mxi+i=1mlog(qi(θ)){\displaystyle {\begin{aligned}{\mathcal {L}}(x_{1},\ldots ,x_{m};\mu ,d)&=\log({\mathcal {L}}(x_{1},\ldots ,x_{m};\mu ,d))\\&=m\mu \left(-1+{\frac {d-1}{2}}\right)+\log(\mu (2-d))\sum _{i=1}^{m}x_{i}+\sum _{i=1}^{m}\log(q_{i}(\theta ))\end{aligned}}}

where

From the log-likelihood function, thelikelihood equations are,[9]

lμ=m(1+d12)+1μi=1mxid12μ2(2d)2i=1mqi(θ)qi(θ){\displaystyle {\frac {\partial l}{\partial \mu }}=m\left(-1+{\frac {d-1}{2}}\right)+{\frac {1}{\mu }}\sum _{i=1}^{m}x_{i}-{\frac {d-1}{2\mu ^{2}(2-d)^{2}}}\sum _{i=1}^{m}{\frac {q_{i}^{'}(\theta )}{q_{i}(\theta )}}}
ld=mμ2i=1mxi2dd2μ(2d)3i=1mi=1mqi(θ)qi(θ){\displaystyle {\frac {\partial l}{\partial d}}=m{\frac {\mu }{2}}-{\frac {\sum _{i=1}^{m}x_{i}}{2-d}}-{\frac {d}{2\mu (2-d)^{3}}}\sum _{i=1}^{m}\sum _{i=1}^{m}{\frac {q_{i}^{'}(\theta )}{q_{i}(\theta )}}}

Straightforward calculations show that,[9]

i=1mqi(θ~)qi(θ~)=m(x¯(2d))2{\displaystyle \sum _{i=1}^{m}{\frac {q_{i}^{'}({\tilde {\theta }})}{q_{i}({\tilde {\theta }})}}=m({\bar {x}}(2-d))^{2}}

whereθ~=d12x¯(2d)2{\displaystyle {\tilde {\theta }}={\frac {d-1}{2{\bar {x}}(2-d)^{2}}}}

  • It can be shown that thelog-likelihood function is strictly concave in the domain of the parameters. Consequently, the MLE is unique.

The likelihood equation does not always have a solution like as it shows the following proposition,

Proposition:[9] LetX1, ...,Xm come from a generalized Hermite distribution with fixedn. Then the MLEs of the parameters areμ^{\displaystyle {\hat {\mu }}} andd~{\displaystyle {\tilde {d}}} if only ifm(2)/x¯2>1{\displaystyle m^{(2)}/{\bar {x}}^{2}>1}, wherem(2)=i=1nxi(xi1)/n{\displaystyle m^{(2)}=\sum _{i=1}^{n}x_{i}(x_{i}-1)/n} indicates the empirical factorial momement of order 2.

Zero frequency and the mean estimators

[edit]

A usual choice for discrete distributions is the zero relative frequency of the data set which is equated to the probability of zero under the assumed distribution. Observing thatf0=exp((a1+a2)){\displaystyle f_{0}=\exp(-(a_{1}+a_{2}))} andμ=a1+2a2{\displaystyle \mu =a_{1}+2a_{2}}. Following the example of Y. C. Patel (1976) the resulting system of equations,

{x¯=a1+2a2f0=exp((a1+a2)){\displaystyle {\begin{cases}{\bar {x}}=a_{1}+2a_{2}\\f_{0}=\exp(-(a_{1}+a_{2}))\end{cases}}}

We obtain thezero frequency and themean estimatora1 ofa1^{\displaystyle {\hat {a_{1}}}} anda2 ofa2^{\displaystyle {\hat {a_{2}}}},[6]

a1^=(x¯+2log(f0)){\displaystyle {\hat {a_{1}}}=-({\bar {x}}+2\log(f_{0}))}
a2^=x¯+log(f0){\displaystyle {\hat {a_{2}}}={\bar {x}}+\log(f_{0})}

wheref0=n0n{\displaystyle f_{0}={\frac {n_{0}}{n}}}, is the zero relative frequency, n > 0

It can be seen that for distributions with a high probability at 0, the efficiency is high.

log(n0n)<x¯<2log(n0n){\displaystyle -\log \left({\frac {n_{0}}{n}}\right)<{\bar {x}}<-2\log \left({\frac {n_{0}}{n}}\right)}

Testing Poisson assumption

[edit]

When Hermite distribution is used to model a data sample is important to check if thePoisson distribution is enough to fit the data. Following the parametrizedprobability mass function used to calculate the maximum likelihood estimator, is important to corroborate the following hypothesis,

{H0:d=1H1:d>1{\displaystyle {\begin{cases}H_{0}:d=1\\H_{1}:d>1\end{cases}}}

Likelihood-ratio test

[edit]

Thelikelihood-ratio test statistic[9] for hermite distribution is,

W=2(L(X;μ^,d^)L(X;μ^,1)){\displaystyle W=2({\mathcal {L}}(X;{\hat {\mu }},{\hat {d}})-{\mathcal {L}}(X;{\hat {\mu }},1))}

WhereL(){\displaystyle {\mathcal {L}}()} is the log-likelihood function. Asd = 1 belongs to the boundary of the domain of parameters, under the null hypothesis,W does not have an asymptoticχ12{\displaystyle \chi _{1}^{2}} distribution as expected. It can be established that the asymptotic distribution ofW is a 50:50 mixture of the constant 0 and theχ12{\displaystyle \chi _{1}^{2}}. The α upper-tail percentage points for this mixture are the same as the 2α upper-tail percentage points for aχ12{\displaystyle \chi _{1}^{2}}; for instance, for α = 0.01, 0.05, and 0.10 they are 5.41189, 2.70554 and 1.64237.

The "score" or Lagrange multiplier test

[edit]

The score statistic is,[9]

S2=2m[m(2)x¯22x¯]2=m(d~1)22{\displaystyle S_{2}=2m\left[{\frac {m^{(2)}-{\bar {x}}^{2}}{2{\bar {x}}}}\right]^{2}={\frac {m({\tilde {d}}-1)^{2}}{2}}}

wherem is the number of observations.

The asymptotic distribution of the score test statistic under the null hypothesis is aχ12{\displaystyle \chi _{1}^{2}} distribution. It may be convenient to use a signed version of the score test, that is,sgn(m(2)x¯2)S{\displaystyle \operatorname {sgn} (m^{(2)}-{\bar {x}}^{2}){\sqrt {S}}}, following asymptotically a standard normal.

See also

[edit]

References

[edit]
  1. ^abcKemp, C.D.;Kemp, A.W. (1965). "Some Properties of the "Hermite" Distribution".Biometrika.52 (3–4):381–394.doi:10.1093/biomet/52.3-4.381.
  2. ^abMcKendrick, A.G. (1926)."Applications of Mathematics to Medical Problems".Proceedings of the Edinburgh Mathematical Society.44:98–130.doi:10.1017/s0013091500034428.
  3. ^Huiming, Zhang; Yunxiao Liu; Bo Li (2014). "Notes on discrete compound Poisson model with applications to risk theory".Insurance: Mathematics and Economics.59:325–336.doi:10.1016/j.insmatheco.2014.09.012.
  4. ^abcdJohnson, N.L., Kemp, A.W., and Kotz, S. (2005) Univariate Discrete Distributions, 3rd Edition, Wiley,ISBN 978-0-471-27246-5.
  5. ^Kemp, ADRIENNE W.; Kemp C.D (1966). "An alternative derivation of the Hermite distribution".Biometrika.53 (3–4):627–628.doi:10.1093/biomet/53.3-4.627.
  6. ^abcPatel, Y.C (1976). "Even Point Estimation and Moment Estimation in Hermite Distribution".Biometrics.32 (4):865–873.doi:10.2307/2529270.JSTOR 2529270.
  7. ^Gupta, R.P.; Jain, G.C. (1974). "A Generalized Hermite distribution and Its Properties".SIAM Journal on Applied Mathematics.27 (2):359–363.doi:10.1137/0127027.JSTOR 2100572.
  8. ^abKotz, Samuel (1982–1989).Encyclopedia of statistical sciences. John Wiley.ISBN 978-0471055525.
  9. ^abcdefghPuig, P. (2003). "Characterizing Additively Closed Discrete Models by a Property of Their Maximum Likelihood Estimators, with an Application to Generalized Hermite Distributions".Journal of the American Statistical Association.98 (463):687–692.doi:10.1198/016214503000000594.JSTOR 30045296.S2CID 120484966.
Discrete
univariate
with finite
support
with infinite
support
Continuous
univariate
supported on a
bounded interval
supported on a
semi-infinite
interval
supported
on the whole
real line
with support
whose type varies
Mixed
univariate
continuous-
discrete
Multivariate
(joint)
Directional
Degenerate
andsingular
Degenerate
Dirac delta function
Singular
Cantor
Families
Retrieved from "https://en.wikipedia.org/w/index.php?title=Hermite_distribution&oldid=1296250658"
Category:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp