Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Negative binomial distribution

From Wikipedia, the free encyclopedia
Probability distribution
Different texts (and even different parts of this article) adopt slightly different definitions for the negative binomial distribution. They can be distinguished by whether the support starts atk = 0 or atk = r, whetherp denotes the probability of a success or of a failure, and whetherr represents success or failure,[1] so identifying the specific parametrization used is crucial in any given text.
Probability mass function

The orange line represents the mean, which is equal to 10 in each of these plots; the green line shows the standard deviation.
NotationNB(r,p){\displaystyle \mathrm {NB} (r,\,p)}
Parametersr > 0 — number of successes until the experiment is stopped (integer, but the definition can also be extended toreals)
p ∈ [0,1] — success probability in each experiment (real)
Supportk ∈ { 0, 1, 2, 3, … } — number of failures
PMFk(k+r1k)(1p)kpr,{\displaystyle k\mapsto {k+r-1 \choose k}\cdot (1-p)^{k}p^{r},} involving abinomial coefficient
CDFkIp(r,k+1),{\displaystyle k\mapsto I_{p}(r,\,k+1),} theregularized incomplete beta function
Meanr(1p)p{\displaystyle {\frac {r(1-p)}{p}}}
Mode{(r1)(1p)pif r>10if r1{\displaystyle {\begin{cases}\left\lfloor {\frac {(r-1)(1-p)}{p}}\right\rfloor &{\text{if }}r>1\\0&{\text{if }}r\leq 1\end{cases}}}
Variancer(1p)p2{\displaystyle {\frac {r(1-p)}{p^{2}}}}
Skewness2p(1p)r{\displaystyle {\frac {2-p}{\sqrt {(1-p)r}}}}
Excess kurtosis6r+p2(1p)r{\displaystyle {\frac {6}{r}}+{\frac {p^{2}}{(1-p)r}}}
MGF(p1(1p)et)r for t<log(1p){\displaystyle {\biggl (}{\frac {p}{1-(1-p)e^{t}}}{\biggr )}^{\!r}{\text{ for }}t<-\log(1-p)}
CF(p1(1p)eit)r with tR{\displaystyle {\biggl (}{\frac {p}{1-(1-p)e^{i\,t}}}{\biggr )}^{\!r}{\text{ with }}t\in \mathbb {R} }
PGF(p1(1p)z)r for |z|<1p{\displaystyle {\biggl (}{\frac {p}{1-(1-p)z}}{\biggr )}^{\!r}{\text{ for }}|z|<{\frac {1}{p}}}
Fisher informationrp2(1p){\displaystyle {\frac {r}{p^{2}(1-p)}}}
Method of momentsr=E[X]2V[X]E[X]{\displaystyle r={\frac {E[X]^{2}}{V[X]-E[X]}}}
p=E[X]V[X]{\displaystyle p={\frac {E[X]}{V[X]}}}

Inprobability theory andstatistics, thenegative binomial distribution, also called aPascal distribution,[2] is adiscrete probability distribution that models the number of failures in a sequence of independent and identically distributedBernoulli trials before a specified/constant/fixed number of successesr{\displaystyle r} occur.[3] For example, we can define rolling a 6 on some dice as a success, and rolling any other number as a failure, and ask how many failure rolls will occur before we see the third success (r=3{\displaystyle r=3}). In such a case, the probability distribution of the number of failures that appear will be a negative binomial distribution.

An alternative formulation is to model the number of total trials (instead of the number of failures). In fact, for a specified (non-random) number of successes(r), the number of failures(nr) is random because the number of total trials(n) is random. For example, we could use the negative binomial distribution to model the number of daysn (random) a certain machine works (specified byr) before it breaks down.

The negative binomial distribution has a varianceμ/p{\displaystyle \mu /p}, with the distribution becoming identical to Poisson in the limitp1{\displaystyle p\to 1} for a given meanμ{\displaystyle \mu } (i.e. when the failures are increasingly rare). Herep[0,1]{\displaystyle p\in [0,1]} is the success probability of each Bernoulli trial. This can make the distribution a usefuloverdispersed alternative to the Poisson distribution, for example for arobust modification ofPoisson regression. In epidemiology, it has been used to model disease transmission for infectious diseases where the likely number of onward infections may vary considerably from individual to individual and from setting to setting.[4] More generally, it may be appropriate where events have positively correlated occurrences causing a largervariance than if the occurrences were independent, due to a positivecovariance term.

The term "negative binomial" is likely due to the fact that a certainbinomial coefficient that appears in the formula for theprobability mass function of the distribution can be written more simply with negative numbers.[5]

Definitions

[edit]

Imagine a sequence of independentBernoulli trials: each trial has two potential outcomes called "success" and "failure." In each trial the probability of success isp{\displaystyle p} and of failure is1p{\displaystyle 1-p}. We observe this sequence until a predefined numberr{\displaystyle r} of successes occurs. Then the random number of observed failures,X{\displaystyle X}, follows thenegative binomial distribution:XNB(r,p){\displaystyle X\sim \operatorname {NB} (r,p)}

Probability mass function

[edit]

Theprobability mass function of the negative binomial distribution isf(k;r,p)Pr(X=k)=(k+r1k)(1p)kpr{\displaystyle f(k;r,p)\equiv \Pr(X=k)={\binom {k+r-1}{k}}(1-p)^{k}p^{r}}wherer is the number of successes,k is the number of failures, andp is the probability of success on each trial.

Here, the quantity in parentheses is thebinomial coefficient, and is equal to(k+r1k)=(k+r1)!(r1)!(k)!=(k+r1)(k+r2)(r)k!=Γ(k+r)k! Γ(r)=((rk)).{\displaystyle {\binom {k+r-1}{k}}={\frac {(k+r-1)!}{(r-1)!\,(k)!}}={\frac {(k+r-1)(k+r-2)\dotsm (r)}{k!}}={\frac {\Gamma (k+r)}{k!\ \Gamma (r)}}=\left(\!\!{r \choose k}\!\!\right).}Note thatΓ(r) is theGamma function, and((rk)){\displaystyle \textstyle \left(\!\!{r \choose k}\!\!\right)} is themultiset coefficient.

There arek failures chosen fromk +r − 1 trials rather thank +r because the last of thek +r trials is by definition a success.

This quantity can alternatively be written in the following manner, explaining the name "negative binomial":

(k+r1)(r)k!=(1)k(r)(r1)(r2)(rk+1)k factorsk!=(1)k(rk).{\displaystyle {\begin{aligned}&{\frac {(k+r-1)\dotsm (r)}{k!}}\\[10pt]={}&(-1)^{k}{\frac {\overbrace {(-r)(-r-1)(-r-2)\dotsm (-r-k+1)} ^{k{\text{ factors}}}}{k!}}=(-1)^{k}{\binom {-r}{{\phantom {-}}k}}.\end{aligned}}}

Note that by the last expression and thebinomial series, for every0 ≤p < 1 andq=1p{\displaystyle q=1-p},

pr=(1q)r=k=0(rk)(q)k=k=0(k+r1k)qk{\displaystyle p^{-r}=(1-q)^{-r}=\sum _{k=0}^{\infty }{\binom {-r}{{\phantom {-}}k}}(-q)^{k}=\sum _{k=0}^{\infty }{\binom {k+r-1}{k}}q^{k}}

hence the terms of the probability mass function indeed add up to one as below.k=0(k+r1k)(1p)kpr=prpr=1{\displaystyle \sum _{k=0}^{\infty }{\binom {k+r-1}{k}}\left(1-p\right)^{k}p^{r}=p^{-r}p^{r}=1}

To understand the above definition of the probability mass function, note that the probability for every specific sequence ofr successes andk failures ispr(1 −p)k, because the outcomes of thek +r trials are supposed to happenindependently. Since ther-th success always comes last, it remains to choose thek trials with failures out of the remainingk +r − 1 trials. The above binomial coefficient, due to its combinatorial interpretation, gives precisely the number of all these sequences of lengthk +r − 1.

An alternate interpretation of the probability mass function's binomial coefficient arises when considering the equivalent multiset coefficient((rk)){\displaystyle \textstyle \left(\!\!{r \choose k}\!\!\right)}. A sequence of trials ending inr successes can be represented by a tuple ofr non-negative integers, where each integer represents the number of failures seen before the next success. Then by applyingstars and bars, it can be seen that the number of such tuples that sum tok (and hence representk total failures andr total successes) is given by((rk)){\displaystyle \textstyle \left(\!\!{r \choose k}\!\!\right)}.

Cumulative distribution function

[edit]

Thecumulative distribution function can be expressed in terms of theregularized incomplete beta function:[3][6]F(k;r,p)Pr(Xk)=Ip(r,k+1).{\displaystyle F(k;r,p)\equiv \Pr(X\leq k)=I_{p}(r,k+1).}(This formula is using the same parameterization as in the article's table, withr the number of successes, andp=r/(r+μ){\displaystyle p=r/(r+\mu )} withμ{\displaystyle \mu } the mean.)

It can also be expressed in terms of thecumulative distribution function of thebinomial distribution:[7]F(k;r,p)=Fbinomial(k;n=k+r,1p).{\displaystyle F(k;r,p)=F_{\text{binomial}}(k;n=k+r,1-p).}

Alternative formulations

[edit]

Some sources may define the negative binomial distribution slightly differently from the primary one here. The most common variations are where the random variableX is counting different things. These variations can be seen in the table here:

X is counting...Probability mass functionFormulaAlternate formula

(using equivalent binomial)

Alternate formula

(simplified using:n=k+r{\textstyle n=k+r})

Support
1k failures, givenr successesf(k;r,p)Pr(X=k)={\textstyle f(k;r,p)\equiv \Pr(X=k)=}(k+r1k)pr(1p)k{\textstyle {\binom {k+r-1}{k}}p^{r}(1-p)^{k}}[8][6][9](k+r1r1)pr(1p)k{\textstyle {\binom {k+r-1}{r-1}}p^{r}(1-p)^{k}}[3][10][11][12](n1k)pr(1p)k{\textstyle {\binom {n-1}{k}}p^{r}(1-p)^{k}}for k=0,1,2,{\displaystyle {\text{for }}k=0,1,2,\ldots }
2n trials, givenr successesf(n;r,p)Pr(X=n)={\textstyle f(n;r,p)\equiv \Pr(X=n)=}(n1r1)pr(1p)nr{\textstyle {\binom {n-1}{r-1}}p^{r}(1-p)^{n-r}}[6][12][13][14][15](n1nr)pr(1p)nr{\textstyle {\binom {n-1}{n-r}}p^{r}(1-p)^{n-r}}for n=r,r+1,r+2,{\displaystyle {\text{for }}n=r,r+1,r+2,\dotsc }
3n trials, givenr failuresf(n;r,p)Pr(X=n)={\textstyle f(n;r,p)\equiv \Pr(X=n)=}(n1r1)pnr(1p)r{\textstyle {\binom {n-1}{r-1}}p^{n-r}(1-p)^{r}}(n1nr)pnr(1p)r{\textstyle {\binom {n-1}{n-r}}p^{n-r}(1-p)^{r}}(n1k)pk(1p)r{\textstyle {\binom {n-1}{k}}p^{k}(1-p)^{r}}
4k successes, givenr failuresf(k;r,p)Pr(X=k)={\textstyle f(k;r,p)\equiv \Pr(X=k)=}(k+r1k)pk(1p)r{\textstyle {\binom {k+r-1}{k}}p^{k}(1-p)^{r}}(k+r1r1)pk(1p)r{\textstyle {\binom {k+r-1}{r-1}}p^{k}(1-p)^{r}}for k=0,1,2,{\displaystyle {\text{for }}k=0,1,2,\ldots }
-k successes, givenn trialsf(k;n,p)Pr(X=k)={\textstyle f(k;n,p)\equiv \Pr(X=k)=}This is thebinomial distribution not the negative binomial:(nk)pk(1p)nk=(nnk)pk(1p)nk=(nk)pk(1p)r{\textstyle {\binom {n}{k}}p^{k}(1-p)^{n-k}={\binom {n}{n-k}}p^{k}(1-p)^{n-k}={\binom {n}{k}}p^{k}(1-p)^{r}}for k=0,1,2,,n{\displaystyle {\text{for }}k=0,1,2,\dotsc ,n}

Each of the four definitions of the negative binomial distribution can be expressed in slightly different but equivalent ways. The first alternative formulation is simply an equivalent form of the binomial coefficient, that is:(ab)=(aab)for  0ba{\textstyle {\binom {a}{b}}={\binom {a}{a-b}}\quad {\text{for }}\ 0\leq b\leq a}. The second alternate formulation somewhat simplifies the expression by recognizing that the total number of trials is simply the number of successes and failures, that is:n=r+k{\textstyle n=r+k}. These second formulations may be more intuitive to understand, however they are perhaps less practical as they have more terms.

  • The definition whereX is the number ofntrials that occur for a given number ofrsuccesses is similar to the primary definition, except that the number of trials is given instead of the number of failures. This addsr to the value of the random variable, shifting its support and mean.
  • The definition whereX is the number ofksuccesses (orntrials) that occur for a given number ofrfailures is similar to the primary definition used in this article, except that numbers of failures and successes are switched when considering what is being counted and what is given. Note however, thatp still refers to the probability of "success".
  • The definition of the negative binomial distribution can be extended to the case where the parameterr can take on a positivereal value. Although it is impossible to visualize a non-integer number of "failures", we can still formally define the distribution through its probability mass function. The problem of extending the definition to real-valued (positive)r boils down to extending the binomial coefficient to its real-valued counterpart, based on thegamma function:(k+r1k)=(k+r1)(k+r2)(r)k!=Γ(k+r)k!Γ(r){\displaystyle {\binom {k+r-1}{k}}={\frac {(k+r-1)(k+r-2)\dotsm (r)}{k!}}={\frac {\Gamma (k+r)}{k!\,\Gamma (r)}}} After substituting this expression in the original definition, we say thatX has a negative binomial (orPólya) distribution if it has aprobability mass function:f(k;r,p)Pr(X=k)=Γ(k+r)k!Γ(r)(1p)kprfor k=0,1,2,{\displaystyle f(k;r,p)\equiv \Pr(X=k)={\frac {\Gamma (k+r)}{k!\,\Gamma (r)}}(1-p)^{k}p^{r}\quad {\text{for }}k=0,1,2,\dotsc } Herer is a real, positive number.

In negative binomial regression,[16] the distribution is specified in terms of its mean,m=r(1p)p{\textstyle m={\frac {r(1-p)}{p}}}, which is then related to explanatory variables as inlinear regression or othergeneralized linear models. From the expression for the meanm, one can derivep=rm+r{\textstyle p={\frac {r}{m+r}}} and1p=mm+r{\textstyle 1-p={\frac {m}{m+r}}}. Then, substituting these expressions inthe one for the probability mass function whenr is real-valued, yields this parametrization of the probability mass function in terms of m:

Pr(X=k)=Γ(r+k)k!Γ(r)(rr+m)r(mr+m)kfor k=0,1,2,{\displaystyle \Pr(X=k)={\frac {\Gamma (r+k)}{k!\,\Gamma (r)}}\left({\frac {r}{r+m}}\right)^{r}\left({\frac {m}{r+m}}\right)^{k}\quad {\text{for }}k=0,1,2,\dotsc }The variance can then be written asm+m2r{\textstyle m+{\frac {m^{2}}{r}}}. Some authors prefer to setα=1r{\textstyle \alpha ={\frac {1}{r}}}, and express the variance asm+αm2{\textstyle m+\alpha m^{2}}. In this context, and depending on the author, either the parameterr or its reciprocalα is referred to as the "dispersion parameter", "shape parameter" or "clustering coefficient",[17] or the "heterogeneity"[16] or "aggregation" parameter.[11] The term "aggregation" is particularly used in ecology when describing counts of individual organisms. Decrease of the aggregation parameterr towards zero corresponds to increasing aggregation of the organisms; increase ofr towards infinity corresponds to absence of aggregation, as can be described byPoisson regression.

Alternative parameterizations

[edit]

Sometimes the distribution is parameterized in terms of its meanμ and varianceσ2:p=μσ2,r=μ2σ2μ,Pr(X=k)=(k+μ2σ2μ1k)(1μσ2)k(μσ2)μ2/(σ2μ)E(X)=μVar(X)=σ2.{\displaystyle {\begin{aligned}&p={\frac {\mu }{\sigma ^{2}}},\\[6pt]&r={\frac {\mu ^{2}}{\sigma ^{2}-\mu }},\\[3pt]&\Pr(X=k)={k+{\frac {\mu ^{2}}{\sigma ^{2}-\mu }}-1 \choose k}\left(1-{\frac {\mu }{\sigma ^{2}}}\right)^{k}\left({\frac {\mu }{\sigma ^{2}}}\right)^{\mu ^{2}/(\sigma ^{2}-\mu )}\\&\operatorname {E} (X)=\mu \\&\operatorname {Var} (X)=\sigma ^{2}.\end{aligned}}}

Another popular parameterization usesr and the failureoddsβ:p=11+βPr(X=k)=(k+r1k)(β1+β)k(11+β)rE(X)=rβVar(X)=rβ(1+β).{\displaystyle {\begin{aligned}&p={\frac {1}{1+\beta }}\\&\Pr(X=k)={k+r-1 \choose k}\left({\frac {\beta }{1+\beta }}\right)^{k}\left({\frac {1}{1+\beta }}\right)^{r}\\&\operatorname {E} (X)=r\beta \\&\operatorname {Var} (X)=r\beta (1+\beta ).\end{aligned}}}

Examples

[edit]

Length of hospital stay

[edit]

Hospitallength of stay is an example of real-world data that can be modelled well with a negative binomial distribution vianegative binomial regression.[18][19]

Selling candy

[edit]

Pat Collis is required to sell candy bars to raise money for the 6th grade field trip. Pat is (somewhat harshly) not supposed to return home until five candy bars have been sold. So the child goes door to door, selling candy bars. At each house, there is a 0.6 probability of selling one candy bar and a 0.4 probability of selling nothing.

What's the probability of selling the last candy bar at then-thhouse?

Successfully selling candy enough times is what defines our stopping criterion (as opposed to failing to sell it), sok in this case represents the number of failures andr represents the number of successes. Recall that theNB(r,p) distribution describes the probability ofk failures andr successes ink +rBernoulli(p) trials with success on the last trial. Selling five candy bars means getting five successes. The number of trials (i.e. houses) this takes is thereforek + 5 =n. The random variable we are interested in is the number of houses, so we substitutek =n − 5 into aNB(5, 0.4) mass function and obtain the following mass function of the distribution of houses (forn ≥ 5):

f(n)=((n5)+51n5)(10.4)50.4n5=(n1n5)352n55n.{\displaystyle f(n)={\binom {(n-5)+5-1}{n-5}}\;(1-0.4)^{5}\;0.4^{n-5}={n-1 \choose n-5}\;3^{5}\;{\frac {2^{n-5}}{5^{n}}}.}

What's the probability that Pat finishes on the tenth house?

f(10)=97977697656250.10033.{\displaystyle f(10)={\frac {979776}{9765625}}\approx 0.10033.\,}

What's the probability that Pat finishes on or before reaching the eighth house?

To finish on or before the eighth house, Pat must finish at the fifth, sixth, seventh, or eighth house. Sum those probabilities:f(5)=24331250.07776f(6)=48631250.15552f(7)=2916156250.18662f(8)=13608781250.17418{\displaystyle {\begin{aligned}f(5)&={\frac {243}{3125}}\approx 0.07776\\f(6)&={\frac {486}{3125}}\approx 0.15552\\f(7)&={\frac {2916}{15625}}\approx 0.18662\\f(8)&={\frac {13608}{78125}}\approx 0.17418\end{aligned}}}j=58f(j)=46413781250.59409.{\displaystyle \sum _{j=5}^{8}f(j)={\frac {46413}{78125}}\approx 0.59409.}

What's the probability that Pat exhausts all 30 houses that happen to stand in the neighborhood?

This can be expressed as the probability that Patdoes not finish on the fifth through the thirtieth house:1j=530f(j)=1I0.4(5,305+1)10.999999823=0.000000177.{\displaystyle 1-\sum _{j=5}^{30}f(j)=1-I_{0.4}(5,30-5+1)\approx 1-0.999999823=0.000000177.}

Because of the rather high probability that Pat will sell to each house (60 percent), the probability of hernot fulfilling her quest is vanishingly slim.

Properties

[edit]

Expectation

[edit]

The expected total number of trials needed to seer successes isrp{\displaystyle {\frac {r}{p}}}. Thus, the expected number offailures would be this value, minus the successes:E[NB(r,p)]=rpr=r(1p)p{\displaystyle \operatorname {E} [\operatorname {NB} (r,p)]={\frac {r}{p}}-r={\frac {r(1-p)}{p}}}

Expectation of successes

[edit]

The expected total number of failures in a negative binomial distribution with parameters(r,p) isr(1 −p)/p. To see this, imagine an experiment simulating the negative binomial is performed many times. That is, a set of trials is performed untilr successes are obtained, then another set of trials, and then another etc. Write down the number of trials performed in each experiment:a,b,c, ... and seta +b +c + ... =N. Now we would expect aboutNp successes in total. Say the experiment was performedn times. Then there arenr successes in total. So we would expectnr =Np, soN/n =r/p. See thatN/n is just the average number of trials per experiment. That is what we mean by "expectation". The average number of failures per experiment isN/nr =r/pr =r(1 −p)/p. This agrees with the mean given in the box on the right-hand side of this page.

A rigorous derivation can be done by representing the negative binomial distribution as the sum of waiting times. LetXrNB(r,p){\displaystyle X_{r}\sim \operatorname {NB} (r,p)} with the conventionX{\displaystyle X} represents the number of failures observed beforer{\displaystyle r} successes with the probability of success beingp{\displaystyle p}. And letYiGeom(p){\displaystyle Y_{i}\sim \mathrm {Geom} (p)} whereYi{\displaystyle Y_{i}} represents the number of failures before seeing a success. We can think ofYi{\displaystyle Y_{i}} as the waiting time (number of failures) between thei{\displaystyle i}th and(i1){\displaystyle (i-1)}th success. ThusXr=Y1+Y2++Yr.{\displaystyle X_{r}=Y_{1}+Y_{2}+\cdots +Y_{r}.}The mean isE[Xr]=E[Y1]+E[Y2]++E[Yr]=r(1p)p,{\displaystyle \operatorname {E} [X_{r}]=\operatorname {E} [Y_{1}]+\operatorname {E} [Y_{2}]+\cdots +\operatorname {E} [Y_{r}]={\frac {r(1-p)}{p}},}which follows from the factE[Yi]=(1p)/p{\displaystyle \operatorname {E} [Y_{i}]=(1-p)/p}.

Variance

[edit]

When counting the number of failures before ther-th success, the variance is r(1 −p)/p2. When counting the number of successes before ther-th failure, as in alternative formulation (3) above, the variance is rp/(1 −p)2.

Relation to the binomial theorem

[edit]

SupposeY is a random variable with abinomial distribution with parametersn andp. Assumep +q = 1, withp,q ≥ 0, then

1=1n=(p+q)n.{\displaystyle 1=1^{n}=(p+q)^{n}.}

UsingNewton's binomial theorem, this can equally be written as:

(p+q)n=k=0(nk)pkqnk,{\displaystyle (p+q)^{n}=\sum _{k=0}^{\infty }{\binom {n}{k}}p^{k}q^{n-k},}

in which the upper bound of summation is infinite. In this case, thebinomial coefficient

(nk)=n(n1)(n2)(nk+1)k!.{\displaystyle {\binom {n}{k}}={n(n-1)(n-2)\cdots (n-k+1) \over k!}.}

is defined whenn is a real number, instead of just a positiveinteger. But in our case of the binomial distribution it is zero whenk >n. We can then say, for example

(p+q)8.3=k=0(8.3k)pkq8.3k.{\displaystyle (p+q)^{8.3}=\sum _{k=0}^{\infty }{\binom {8.3}{k}}p^{k}q^{8.3-k}.}

Now supposer > 0 and we use a negative exponent:

1=prpr=pr(1q)r=prk=0(rk)(q)k.{\displaystyle 1=p^{r}\cdot p^{-r}=p^{r}(1-q)^{-r}=p^{r}\sum _{k=0}^{\infty }{\binom {-r}{k}}(-q)^{k}.}

Then all of the terms are positive, and the term

pr(rk)(q)k=(k+r1k)prqk{\displaystyle p^{r}{\binom {-r}{k}}(-q)^{k}={\binom {k+r-1}{k}}p^{r}q^{k}}

is just the probability that the number of failures before ther-th success is equal tok, providedr is an integer. (Ifr is a negative non-integer, so that the exponent is a positive non-integer, then some of the terms in the sum above are negative, so we do not have a probability distribution on the set of all nonnegative integers.)

Now we also allow non-integer values ofr.

Recall from above that

The sum of independent negative-binomially distributed random variablesr1 andr2 with the same value for parameterp is negative-binomially distributed with the samep but withr-value r1 +r2.

This property persists when the definition is thus generalized, and affords a quick way to see that the negative binomial distribution isinfinitely divisible.

Recurrence relations

[edit]

The followingrecurrence relations hold:

For the probability mass function{(k+1)Pr(X=k+1)(1p)Pr(X=k)(k+r)=0,Pr(X=0)=(1p)r.{\displaystyle {\begin{cases}(k+1)\Pr(X=k+1)-(1-p)\Pr(X=k)(k+r)=0,\\[5pt]\Pr(X=0)=(1-p)^{r}.\end{cases}}}

For the momentsmk=E(Xk),{\displaystyle m_{k}=\mathbb {E} (X^{k}),}mk+1=rPmk+(P2+P)dmkdP,P:=(1p)/p,m0=1.{\displaystyle m_{k+1}=rPm_{k}+(P^{2}+P){dm_{k} \over dP},\quad P:=(1-p)/p,\quad m_{0}=1.}

For the cumulantsκk+1=(Q1)QdκkdQ,Q:=1/p,κ1=r(Q1).{\displaystyle \kappa _{k+1}=(Q-1)Q{d\kappa _{k} \over dQ},\quad Q:=1/p,\quad \kappa _{1}=r(Q-1).}

Related distributions

[edit]

Poisson distribution

[edit]

Consider a sequence of negative binomial random variables where the stopping parameterr goes to infinity, while the probabilityp of success in each trial goes to one, in such a way as to keep the mean of the distribution (i.e. the expected number of failures) constant. Denoting this mean asλ, the parameterp will bep =r/(r +λ)Mean:λ=(1p)rpp=rr+λ,Variance:λ(1+λr)>λ,thus always overdispersed.{\displaystyle {\begin{aligned}{\text{Mean:}}\quad &\lambda ={\frac {(1-p)r}{p}}\quad \Rightarrow \quad p={\frac {r}{r+\lambda }},\\{\text{Variance:}}\quad &\lambda \left(1+{\frac {\lambda }{r}}\right)>\lambda ,\quad {\text{thus always overdispersed}}.\end{aligned}}}

Under this parametrization the probability mass function will bef(k;r,p)=Γ(k+r)k!Γ(r)(1p)kpr=λkk!Γ(r+k)Γ(r)(r+λ)k1(1+λr)r{\displaystyle f(k;r,p)={\frac {\Gamma (k+r)}{k!\cdot \Gamma (r)}}(1-p)^{k}p^{r}={\frac {\lambda ^{k}}{k!}}\cdot {\frac {\Gamma (r+k)}{\Gamma (r)\;(r+\lambda )^{k}}}\cdot {\frac {1}{\left(1+{\frac {\lambda }{r}}\right)^{r}}}}

Now if we consider the limit asr → ∞, the second factor will converge to one, and the third to the exponent function:limrf(k;r,p)=λkk!11eλ,{\displaystyle \lim _{r\to \infty }f(k;r,p)={\frac {\lambda ^{k}}{k!}}\cdot 1\cdot {\frac {1}{e^{\lambda }}},}which is the mass function of aPoisson-distributed random variable with expected value λ.

In other words, the alternatively parameterized negative binomial distributionconverges to the Poisson distribution andr controls the deviation from the Poisson. This makes the negative binomial distribution suitable as a robust alternative to the Poisson, which approaches the Poisson for larger, but which has larger variance than the Poisson for smallr.Poisson(λ)=limrNB(r,rr+λ).{\displaystyle \operatorname {Poisson} (\lambda )=\lim _{r\to \infty }\operatorname {NB} \left(r,{\frac {r}{r+\lambda }}\right).}

Gamma–Poisson mixture

[edit]

The negative binomial distribution also arises as a continuous mixture ofPoisson distributions (i.e. acompound probability distribution) where the mixing distribution of the Poisson rate is agamma distribution. That is, we can view the negative binomial as aPoisson(λ) distribution, whereλ is itself a random variable, distributed as a gamma distribution with shaper and scaleθ = (1 −p)/p or correspondingly rateβ =p/(1 −p).

To display the intuition behind this statement, consider two independent Poisson processes, "Success" and "Failure", with intensitiesp and1 −p. Together, the Success and Failure processes are equivalent to a single Poisson process of intensity 1, where an occurrence of the process is a success if a corresponding independent coin toss comes up heads with probabilityp; otherwise, it is a failure. Ifr is a counting number, the coin tosses show that the count of successes before ther-th failure follows a negative binomial distribution with parametersr and1 −p. The count is also, however, the count of the Success Poisson process at the random timeT of ther-th occurrence in the Failure Poisson process. The Success count follows a Poisson distribution with meanpT, whereT is the waiting time forr occurrences in a Poisson process of intensity1 −p, i.e.,T is gamma-distributed with shape parameterr and intensity1 −p. Thus, the negative binomial distribution is equivalent to a Poisson distribution with meanpT, where the random variateT is gamma-distributed with shape parameterr and intensity(1 −p). The preceding paragraph follows, becauseλ =pT is gamma-distributed with shape parameterr and intensity(1 −p)/p.

The following formal derivation (which does not depend onr being a counting number) confirms the intuition.

0fPoisson(λ)(k)×fGamma(r,p1p)(λ)dλ=0λkk!eλ×1Γ(r)(p1pλ)r1ep1pλ(p1p)dλ=(p1p)r1k!Γ(r)0λr+k1eλp+1p1pdλ=(p1p)r1k!Γ(r)Γ(r+k)(1p)k+r0fGamma(k+r,11p)(λ)dλ=Γ(r+k)k!Γ(r)(1p)kpr=f(k;r,p).{\displaystyle {\begin{aligned}&\int _{0}^{\infty }f_{\operatorname {Poisson} (\lambda )}(k)\times f_{\operatorname {Gamma} \left(r,\,{\frac {p}{1-p}}\right)}(\lambda )\,\mathrm {d} \lambda \\[8pt]={}&\int _{0}^{\infty }{\frac {\lambda ^{k}}{k!}}e^{-\lambda }\times {\frac {1}{\Gamma (r)}}\left({\frac {p}{1-p}}\lambda \right)^{r-1}e^{-{\frac {p}{1-p}}\lambda }\,\left({\frac {p}{1-p}}\,\right)\mathrm {d} \lambda \\[8pt]={}&\left({\frac {p}{1-p}}\right)^{r}{\frac {1}{k!\,\Gamma (r)}}\int _{0}^{\infty }\lambda ^{r+k-1}e^{-\lambda {\frac {p+1-p}{1-p}}}\;\mathrm {d} \lambda \\[8pt]={}&\left({\frac {p}{1-p}}\right)^{r}{\frac {1}{k!\,\Gamma (r)}}\Gamma (r+k)(1-p)^{k+r}\int _{0}^{\infty }f_{\operatorname {Gamma} \left(k+r,{\frac {1}{1-p}}\right)}(\lambda )\;\mathrm {d} \lambda \\[8pt]={}&{\frac {\Gamma (r+k)}{k!\;\Gamma (r)}}\;(1-p)^{k}\,p^{r}\\[8pt]={}&f(k;r,p).\end{aligned}}}

Because of this, the negative binomial distribution is also known as thegamma–Poisson (mixture) distribution. The negative binomial distribution was originally derived as a limiting case of the gamma-Poisson distribution.[20]

Distribution of a sum of geometrically distributed random variables

[edit]

IfYr is a random variable following the negative binomial distribution with parametersr andp, and support{0, 1, 2, ...}, thenYr is a sum ofrindependent variables following thegeometric distribution (on{0, 1, 2, ...}) with parameterp. As a result of thecentral limit theorem,Yr (properly scaled and shifted) is therefore approximatelynormal for sufficiently large r.

Furthermore, ifBs+r is a random variable following thebinomial distribution with parameterss +r andp, then

Pr(Yrs)=1Ip(s+1,r)=1Ip((s+r)(r1),(r1)+1)=1Pr(Bs+rr1)=Pr(Bs+rr)=Pr(after s+r trials, there are at least r successes).{\displaystyle {\begin{aligned}\Pr(Y_{r}\leq s)&{}=1-I_{p}(s+1,r)\\[5pt]&{}=1-I_{p}((s+r)-(r-1),(r-1)+1)\\[5pt]&{}=1-\Pr(B_{s+r}\leq r-1)\\[5pt]&{}=\Pr(B_{s+r}\geq r)\\[5pt]&{}=\Pr({\text{after }}s+r{\text{ trials, there are at least }}r{\text{ successes}}).\end{aligned}}}

In this sense, the negative binomial distribution is the "inverse" of the binomial distribution.

The sum of independent negative-binomially distributed random variablesr1 andr2 with the same value for parameterp is negative-binomially distributed with the samep but withr-value r1 +r2.

The negative binomial distribution isinfinitely divisible, i.e., ifY has a negative binomial distribution, then for any positive integern, there exist independent identically distributed random variablesY1, ...,Yn whose sum has the same distribution thatY has.

Representation as compound Poisson distribution

[edit]

The negative binomial distributionNB(r,p) can be represented as acompound Poisson distribution: Let(Yn)nN{\textstyle (Y_{n})_{n\,\in \,\mathbb {N} }} denote a sequence ofindependent and identically distributed random variables, each one having thelogarithmic series distributionLog(p), with probability mass function

f(k;r,p)=pkkln(1p),kN.{\displaystyle f(k;r,p)={\frac {-p^{k}}{k\ln(1-p)}},\qquad k\in {\mathbb {N} }.}

LetN be a random variable,independent of the sequence, and suppose thatN has aPoisson distribution with meanλ = −r ln(1 −p). Then the random sum

X=n=1NYn{\displaystyle X=\sum _{n=1}^{N}Y_{n}}

isNB(r,p)-distributed. To prove this, we calculate theprobability generating functionGX ofX, which is the composition of the probability generating functionsGN andGY1. Using

GN(z)=exp(λ(z1)),zR,{\displaystyle G_{N}(z)=\exp(\lambda (z-1)),\qquad z\in \mathbb {R} ,}

and

GY1(z)=ln(1pz)ln(1p),|z|<1p,{\displaystyle G_{Y_{1}}(z)={\frac {\ln(1-pz)}{\ln(1-p)}},\qquad |z|<{\frac {1}{p}},}

we obtain

GX(z)=GN(GY1(z))=exp[λ(ln(1pz)ln(1p)1)]=exp[r(ln(1pz)ln(1p))]=(1p1pz)r,|z|<1p,{\displaystyle {\begin{aligned}G_{X}(z)&=G_{N}(G_{Y_{1}}(z))\\[4pt]&=\exp \left[\lambda \left({\frac {\ln(1-pz)}{\ln(1-p)}}-1\right)\right]\\[1ex]&=\exp \left[-r\left(\ln(1-pz)-\ln(1-p)\right)\right]\\[1ex]&=\left({\frac {1-p}{1-pz}}\right)^{r},\qquad |z|<{\frac {1}{p}},\end{aligned}}}

which is the probability generating function of theNB(r,p) distribution.

The following table describes four distributions related to the number of successes in a sequence of draws:

With replacementsNo replacements
Given number of drawsbinomial distributionhypergeometric distribution
Given number of failuresnegative binomial distributionnegative hypergeometric distribution

(a,b,0) class of distributions

[edit]

The negative binomial, along with the Poisson and binomial distributions, is a member of the(a,b, 0) class of distributions. All three of these distributions are special cases of thePanjer distribution. They are also members of anatural exponential family.

Statistical inference

[edit]

Parameter estimation

[edit]

MVUE forp

[edit]

Supposep is unknown and an experiment is conducted where it is decided ahead of time that sampling will continue untilr successes are found. Asufficient statistic for the experiment isk, the number of failures.

In estimatingp, theminimum variance unbiased estimator is

p^=r1r+k1.{\displaystyle {\widehat {p}}={\frac {r-1}{r+k-1}}.}

Maximum likelihood estimation

[edit]

Whenr is known, themaximum likelihood estimate ofp is

p~=rr+k,{\displaystyle {\widetilde {p}}={\frac {r}{r+k}},}

but this is abiased estimate. Its inverse(r +k)/r, is an unbiased estimate of1/p, however.[21]

Whenr is unknown, the maximum likelihood estimator forp andr together only exists for samples for which the sample variance is larger than the sample mean.[22] Thelikelihood function forNiid observations(k1, ...,kN) is

L(r,p)=i=1Nf(ki;r,p){\displaystyle L(r,p)=\prod _{i=1}^{N}f(k_{i};r,p)\,\!}

from which we calculate the log-likelihood function

(r,p)=i=1N[lnΓ(ki+r)ln(ki!)+kiln(1p)]+N[rlnplnΓ(r)].{\displaystyle \ell (r,p)=\sum _{i=1}^{N}\left[\ln \Gamma (k_{i}+r)-\ln(k_{i}!)+k_{i}\ln(1-p)\right]+N\left[r\ln p-\ln \Gamma (r)\right].}

To find the maximum we take the partial derivatives with respect tor andp and set them equal to zero:

(r,p)p=[i=1Nki11p]+Nr1p=0{\displaystyle {\frac {\partial \ell (r,p)}{\partial p}}=-\left[\sum _{i=1}^{N}k_{i}{\frac {1}{1-p}}\right]+Nr{\frac {1}{p}}=0} and

(r,p)r=[i=1Nψ(ki+r)]Nψ(r)+Nln(p)=0{\displaystyle {\frac {\partial \ell (r,p)}{\partial r}}=\left[\sum _{i=1}^{N}\psi (k_{i}+r)\right]-N\psi (r)+N\ln(p)=0}

where

ψ(k)=Γ(k)Γ(k){\displaystyle \psi (k)={\frac {\Gamma '(k)}{\Gamma (k)}}\!} is thedigamma function.

Solving the first equation forp gives:

p=NrNr+i=1Nki{\displaystyle p={\frac {Nr}{Nr+\sum _{i=1}^{N}k_{i}}}}

Substituting this in the second equation gives:

(r,p)r=[i=1Nψ(ki+r)]Nψ(r)+Nln(rr+i=1Nki/N)=0{\displaystyle {\frac {\partial \ell (r,p)}{\partial r}}=\left[\sum _{i=1}^{N}\psi (k_{i}+r)\right]-N\psi (r)+N\ln \left({\frac {r}{r+\sum _{i=1}^{N}k_{i}/N}}\right)=0}

This equation cannot be solved forr inclosed form. If a numerical solution is desired, an iterative technique such asNewton's method can be used. Alternatively, theexpectation–maximization algorithm can be used.[22]

Occurrence and applications

[edit]

Waiting time in a Bernoulli process

[edit]

Letk andr be integers withk non-negative andr positive. In a sequence of independentBernoulli trials with success probabilityp, the negative binomial gives the probability ofk successes andr failures, with a failure on the last trial. Therefore, the negative binomial distribution represents the probability distribution of the number of successes before ther-th failure in aBernoulli process, with probabilityp of successes on each trial.

Consider the following example. Suppose we repeatedly throw a die, and consider a 1 to be a failure. The probability of success on each trial is 5/6. The number of successes before the third failure belongs to the infinite set{ 0, 1, 2, 3, ... }. That number of successes is a negative-binomially distributed random variable.

Whenr = 1 we get the probability distribution of number of successes before the first failure (i.e. the probability of the first failure occurring on the(k + 1)-st trial), which is ageometric distribution:f(k;r,p)=(1p)pk{\displaystyle f(k;r,p)=(1-p)\cdot p^{k}}

Overdispersed Poisson

[edit]

The negative binomial distribution, especially in its alternative parameterization described above, can be used as an alternative to the Poisson distribution. It is especially useful for discrete data over an unbounded positive range whose samplevariance exceeds the samplemean. In such cases, the observations areoverdispersed with respect to a Poisson distribution, for which the mean is equal to the variance. Hence a Poisson distribution is not an appropriate model. Since the negative binomial distribution has one more parameter than the Poisson, the second parameter can be used to adjust the variance independently of the mean. SeeCumulants of some discrete probability distributions.

An application of this is to annual counts oftropical cyclones in theNorth Atlantic or to monthly to 6-monthly counts of wintertimeextratropical cyclones over Europe, for which the variance is greater than the mean.[23][24][25] In the case of modest overdispersion, this may produce substantially similar results to an overdispersed Poisson distribution.[26][27]

Negative binomial modeling is widely employed in ecology and biodiversity research for analyzingcount data where overdispersion is very common. This is because overdispersion is indicative of biological aggregation, such as species or communities forming clusters. Ignoring overdispersion can lead to significantly inflated model parameters, resulting in misleading statistical inferences. The negative binomial distribution effectively addresses overdispersed counts by permitting the variance to vary quadratically with the mean. An additional dispersion parameter governs the slope of the quadratic term, determining the severity of overdispersion. The model's quadratic mean-variance relationship proves to be a realistic approach for handling overdispersion, as supported by empirical evidence from many studies. Overall, the NB model offers two attractive features: (1) the convenient interpretation of the dispersion parameter as an index of clustering or aggregation, and (2) its tractable form, featuring a closed expression for the probability mass function.[28]

In genetics, the negative binomial distribution is commonly used to model data in the form of discrete sequence read counts from high-throughput RNA and DNA sequencing experiments.[29][30][31][32]

In epidemiology of infectious diseases, the negative binomial has been used as a better option than the Poisson distribution to model overdispersed counts of secondary infections from one infected case (super-spreading events).[33]

Multiplicity observations (physics)

[edit]

The negative binomial distribution has been the most effectivestatistical model for a broad range of multiplicity observations inparticle collision experiments, e.g.,pp¯, hh, hA, AA, e+e{\displaystyle p{\bar {p}},\ hh,\ hA,\ AA,\ e^{+}e^{-}}[34][35][36][37][38] (See[39] for an overview), and is argued to be ascale-invariant property of matter,[40][41] providing the best fit for astronomical observations, where it predicts the number of galaxies in a region of space.[42][43][44][45] The phenomenological justification for the effectiveness of the negative binomial distribution in these contexts remained unknown for fifty years, since their first observation in 1973.[46] In 2023, a proof fromfirst principles was eventually demonstrated by Scott V. Tezlaf, where it was shown that the negative binomial distribution emerges fromsymmetries in thedynamical equations of acanonical ensemble of particles inMinkowski space.[47] Roughly, given an expected number of trialsn{\displaystyle \langle n\rangle } and expected number of successesr{\displaystyle \langle r\rangle }, where

nr=k,p=rnn=k1p,r=kp1p,{\displaystyle {\begin{aligned}\langle {\mathcal {n}}\rangle -\langle r\rangle &=k,&\langle p\rangle &={\frac {\langle r\rangle }{\langle {\mathcal {n}}\rangle }}\\[1ex]\implies \langle {\mathcal {n}}\rangle &={\frac {k}{1-\langle p\rangle }},&\langle {r}\rangle &={\frac {k\langle p\rangle }{1-\langle p\rangle }},\end{aligned}}}

anisomorphic set of equations can be identified with the parameters of arelativisticcurrent density of a canonical ensemble of massive particles, via

c2ρ2j2=c2ρ02,βv2=j2c2ρ2c2ρ2=c2ρ021βv2,j2=c2ρ02βv21βv2,{\displaystyle {\begin{aligned}c^{2}\left\langle \rho ^{2}\right\rangle -\left\langle j^{2}\right\rangle &=c^{2}\rho _{0}^{2},&\left\langle \beta _{v}^{2}\right\rangle &={\frac {\left\langle j^{2}\right\rangle }{c^{2}\langle \rho ^{2}\rangle }}\\[1ex]\implies c^{2}\left\langle \rho ^{2}\right\rangle &={\frac {c^{2}\rho _{0}^{2}}{1-\left\langle \beta _{v}^{2}\right\rangle }},&\left\langle j^{2}\right\rangle &={\frac {c^{2}\rho _{0}^{2}\left\langle \beta _{v}^{2}\right\rangle }{1-\left\langle \beta _{v}^{2}\right\rangle }},\end{aligned}}}

whereρ0{\displaystyle \rho _{0}} is the restdensity,ρ2{\displaystyle \langle \rho ^{2}\rangle } is the relativistic mean square density,j2{\displaystyle \langle j^{2}\rangle } is the relativistic mean square current density, andβv2=v2/c2{\displaystyle \langle \beta _{v}^{2}\rangle =\langle v^{2}\rangle /c^{2}}, wherev2{\displaystyle \langle v^{2}\rangle } is themean square speed of the particle ensemble andc{\displaystyle c} is thespeed of light—such that one can establish the followingbijective map:

c2ρ02k,βv2p,c2ρ2n,j2r.{\displaystyle {\begin{aligned}c^{2}\rho _{0}^{2}&\mapsto k,&\langle \beta _{v}^{2}\rangle &\mapsto \langle p\rangle ,\\[1ex]c^{2}\langle \rho ^{2}\rangle &\mapsto \langle {\mathcal {n}}\rangle ,&\langle j^{2}\rangle &\mapsto \langle r\rangle .\end{aligned}}}

A rigorous alternative proof of the above correspondence has also been demonstrated throughquantum mechanics via the Feynmanpath integral.[47]

History

[edit]

This distribution was first studied in 1713 byPierre Remond de Montmort in hisEssay d'analyse sur les jeux de hazard, as the distribution of the number of trials required in an experiment to obtain a given number of successes.[48] It had previously been mentioned byPascal.[49]

See also

[edit]

References

[edit]
  1. ^DeGroot, Morris H. (1986).Probability and Statistics (Second ed.). Addison-Wesley. pp. 258–259.ISBN 0-201-11366-X.LCCN 84006269.OCLC 10605205.
  2. ^Pascal distribution, Univariate Distribution Relationships, Larry Leemis
  3. ^abcWeisstein, Eric."Negative Binomial Distribution".Wolfram MathWorld. Wolfram Research. Retrieved11 October 2020.
  4. ^e.g.Lloyd-Smith, J. O.; Schreiber, S. J.; Kopp, P. E.; Getz, W. M. (2005)."Superspreading and the effect of individual variation on disease emergence".Nature.438 (7066):355–359.Bibcode:2005Natur.438..355L.doi:10.1038/nature04153.PMC 7094981.PMID 16292310.
    The overdispersion parameter is usually denoted by the letterk{\displaystyle k} in epidemiology, rather thanr{\displaystyle r} as here.
  5. ^Casella, George; Berger, Roger L. (2002).Statistical inference (2nd ed.). Thomson Learning. p. 95.ISBN 0-534-24312-6.
  6. ^abcCook, John D."Notes on the Negative Binomial Distribution"(PDF).
  7. ^Morris K W (1963), A note on direct and inverse sampling, Biometrika, 50, 544–545.
  8. ^"Mathworks: Negative Binomial Distribution".
  9. ^Saha, Abhishek."Introduction to Probability / Fundamentals of Probability: Lecture 14"(PDF).
  10. ^SAS Institute, "Negative Binomial Distribution",SAS(R) 9.4 Functions and CALL Routines: Reference, Fourth Edition, SAS Institute, Cary, NC, 2016.
  11. ^abCrawley, Michael J. (2012).The R Book. Wiley.ISBN 978-1-118-44896-0.
  12. ^ab"Set theory: Section 3.2.5 – Negative Binomial Distribution"(PDF).
  13. ^"Randomservices.org, Chapter 10: Bernoulli Trials, Section 4: The Negative Binomial Distribution".
  14. ^"Stat Trek: Negative Binomial Distribution".
  15. ^Wroughton, Jacqueline."Distinguishing Between Binomial, Hypergeometric and Negative Binomial Distributions"(PDF).
  16. ^abHilbe, Joseph M. (2011).Negative Binomial Regression (Second ed.). Cambridge, UK: Cambridge University Press.ISBN 978-0-521-19815-8.
  17. ^Lloyd-Smith, J. O. (2007)."Maximum Likelihood Estimation of the Negative Binomial Dispersion Parameter for Highly Overdispersed Data, with Applications to Infectious Diseases".PLoS ONE.2 (2) e180.Bibcode:2007PLoSO...2..180L.doi:10.1371/journal.pone.0000180.PMC 1791715.PMID 17299582.Open access icon
  18. ^Carter, E.M., Potts, H.W.W. (4 April 2014)."Predicting length of stay from an electronic patient record system: a primary total knee replacement example".BMC Medical Informatics and Decision Making.14 26.doi:10.1186/1472-6947-14-26.PMC 3992140.PMID 24708853.{{cite journal}}: CS1 maint: multiple names: authors list (link)Open access icon
  19. ^Orooji, Arezoo; Nazar, Eisa; Sadeghi, Masoumeh; Moradi, Ali; Jafari, Zahra; Esmaily, Habibollah (2021-04-30)."Factors associated with length of stay in hospital among the elderly patients using count regression models".Medical Journal of the Islamic Republic of Iran.35: 5.doi:10.47176/mjiri.35.5.PMC 8111647.PMID 33996656.
  20. ^Greenwood, M.; Yule, G. U. (1920)."An inquiry into the nature of frequency distributions representative of multiple happenings with particular reference of multiple attacks of disease or of repeated accidents".J R Stat Soc.83 (2):255–279.doi:10.2307/2341080.JSTOR 2341080.
  21. ^Haldane, J. B. S. (1945). "On a Method of Estimating Frequencies".Biometrika.33 (3):222–225.doi:10.1093/biomet/33.3.222.hdl:10338.dmlcz/102575.JSTOR 2332299.PMID 21006837.
  22. ^abAramidis, K. (1999)."An EM algorithm for estimating negative binomial parameters".Australian & New Zealand Journal of Statistics.41 (2):213–221.doi:10.1111/1467-842X.00075.S2CID 118758171.
  23. ^Villarini, G.; Vecchi, G.A.; Smith, J.A. (2010)."Modeling of the dependence of tropical storm counts in the North Atlantic Basin on climate indices".Monthly Weather Review.138 (7):2681–2705.Bibcode:2010MWRv..138.2681V.doi:10.1175/2010MWR3315.1.
  24. ^Mailier, P.J.; Stephenson, D.B.; Ferro, C.A.T.; Hodges, K.I. (2006)."Serial Clustering of Extratropical Cyclones".Monthly Weather Review.134 (8):2224–2240.Bibcode:2006MWRv..134.2224M.doi:10.1175/MWR3160.1.
  25. ^Vitolo, R.; Stephenson, D.B.; Cook, Ian M.; Mitchell-Wallace, K. (2009). "Serial clustering of intense European storms".Meteorologische Zeitschrift.18 (4):411–424.Bibcode:2009MetZe..18..411V.doi:10.1127/0941-2948/2009/0393.S2CID 67845213.
  26. ^McCullagh, Peter;Nelder, John (1989).Generalized Linear Models (Second ed.). Boca Raton: Chapman and Hall/CRC.ISBN 978-0-412-31760-6.
  27. ^Cameron, Adrian C.; Trivedi, Pravin K. (1998).Regression analysis of count data. Cambridge University Press.ISBN 978-0-521-63567-7.
  28. ^Stoklosa, J.; Blakey, R.V.; Hui, F.K.C. (2022)."An Overview of Modern Applications of Negative Binomial Modelling in Ecology and Biodiversity".Diversity.14 (5): 320.Bibcode:2022Diver..14..320S.doi:10.3390/d14050320.
  29. ^Robinson, M.D.; Smyth, G.K. (2007)."Moderated statistical tests for assessing differences in tag abundance".Bioinformatics.23 (21):2881–2887.doi:10.1093/bioinformatics/btm453.PMID 17881408.
  30. ^"Differential analysis of count data – the"(PDF).
  31. ^Airoldi, E. M.; Cohen, W. W.; Fienberg, S. E. (June 2005). "Bayesian Models for Frequent Terms in Text".Proceedings of the Classification Society of North America and INTERFACE Annual Meetings. Vol. 990. St. Louis, MO, USA. p. 991.
  32. ^Chen, Yunshun; Davis, McCarthy (September 25, 2014)."edgeR: differential expression analysis of digital gene expression data"(PDF). RetrievedOctober 14, 2014.
  33. ^Lloyd-Smith, J. O.; Schreiber, S. J.; Kopp, P. E.; Getz, W. M. (2005)."Superspreading and the effect of individual variation on disease emergence".Nature.438 (7066):355–359.Bibcode:2005Natur.438..355L.doi:10.1038/nature04153.PMC 7094981.PMID 16292310.
  34. ^Grosse-Oetringhaus, Jan Fiete; Reygers, Klaus (2010-08-01)."Charged-particle multiplicity in proton–proton collisions".Journal of Physics G: Nuclear and Particle Physics.37 (8) 083001.arXiv:0912.0023.doi:10.1088/0954-3899/37/8/083001.ISSN 0954-3899.S2CID 119233810.
  35. ^Rybczyński, Maciej; Wilk, Grzegorz; Włodarczyk, Zbigniew (2019-05-31)."Intriguing properties of multiplicity distributions".Physical Review D.99 (9) 094045.arXiv:1811.07197.Bibcode:2019PhRvD..99i4045R.doi:10.1103/PhysRevD.99.094045.ISSN 2470-0010.
  36. ^Tarnowsky, Terence J.; Westfall, Gary D. (2013-07-09)."First study of the negative binomial distribution applied to higher moments of net-charge and net-proton multiplicity distributions".Physics Letters B.724 (1):51–55.arXiv:1210.8102.Bibcode:2013PhLB..724...51T.doi:10.1016/j.physletb.2013.05.064.ISSN 0370-2693.
  37. ^Derrick, M.; Gan, K. K.; Kooijman, P.; Loos, J. S.; Musgrave, B.; Price, L. E.; Repond, J.; Schlereth, J.; Sugano, K.; Weiss, J. M.; Wood, D. E.; Baranko, G.; Blockus, D.; Brabson, B.; Brom, J. M. (1986-12-01)."Study of quark fragmentation in ${e}^{+}$${e}^{\mathrm{\ensuremath{-}}}$ annihilation at 29 GeV: Charged-particle multiplicity and single-particle rapidity distributions".Physical Review D.34 (11):3304–3320.doi:10.1103/PhysRevD.34.3304.hdl:1808/15222.PMID 9957066.
  38. ^Zborovský, I. (2018-10-10)."Three-component multiplicity distribution, oscillation of combinants and properties of clans in pp collisions at the LHC".The European Physical Journal C.78 (10) 816.arXiv:1811.11230.Bibcode:2018EPJC...78..816Z.doi:10.1140/epjc/s10052-018-6287-x.ISSN 1434-6052.
  39. ^Kittel, Wolfram; De Wolf, Eddi A (2005).Soft multihardon dynamics. World Scientific.
  40. ^Schaeffer, R (1984). "Determination of the galaxy N-point correlation function".Astronomy and Astrophysics.134 (2): L15.Bibcode:1984A&A...134L..15S.
  41. ^Schaeffer, R (1985). "The probability generating function for galaxy clustering".Astronomy and Astrophysics.144 (1):L1–L4.Bibcode:1985A&A...144L...1S.
  42. ^Perez, Lucia A.; Malhotra, Sangeeta; Rhoads, James E.; Tilvi, Vithal (2021-01-07)."Void Probability Function of Simulated Surveys of High-redshift Ly α Emitters".The Astrophysical Journal.906 (1): 58.arXiv:2011.03556.Bibcode:2021ApJ...906...58P.doi:10.3847/1538-4357/abc88b.ISSN 1538-4357.
  43. ^Hurtado-Gil, Lluís; Martínez, Vicent J.; Arnalte-Mur, Pablo; Pons-Bordería, María-Jesús; Pareja-Flores, Cristóbal; Paredes, Silvestre (2017-05-01)."The best fit for the observed galaxy counts-in-cell distribution function".Astronomy & Astrophysics.601: A40.arXiv:1703.01087.Bibcode:2017A&A...601A..40H.doi:10.1051/0004-6361/201629097.ISSN 0004-6361.
  44. ^Elizalde, E.; Gaztanaga, E. (January 1992)."Void probability as a function of the void's shape and scale-invariant models".Monthly Notices of the Royal Astronomical Society.254 (2):247–256.doi:10.1093/mnras/254.2.247.hdl:2060/19910019799.ISSN 0035-8711.
  45. ^Hameeda, M; Plastino, Angelo; Rocca, M C (2021-03-01)."Generalized Poisson distributions for systems with two-particle interactions".IOP SciNotes.2 (1): 015003.Bibcode:2021IOPSN...2a5003H.doi:10.1088/2633-1357/abec9f.hdl:11336/181371.ISSN 2633-1357.
  46. ^Giovannini, A. (June 1973). ""Thermal chaos" and "coherence" in multiplicity distributions at high energies".Il Nuovo Cimento A.15 (3):543–551.Bibcode:1973NCimA..15..543G.doi:10.1007/bf02734689.ISSN 0369-3546.S2CID 118805136.
  47. ^abTezlaf, Scott V. (2023-09-29)."Significance of the negative binomial distribution in multiplicity phenomena".Physica Scripta.98 (11).arXiv:2310.03776.Bibcode:2023PhyS...98k5310T.doi:10.1088/1402-4896/acfead.ISSN 0031-8949.S2CID 263300385.
  48. ^Montmort PR de (1713) Essai d'analyse sur les jeux de hasard. 2nd ed. Quillau, Paris
  49. ^Pascal B (1679) Varia Opera Mathematica. D. Petri de Fermat. Tolosae
Discrete
univariate
with finite
support
with infinite
support
Continuous
univariate
supported on a
bounded interval
supported on a
semi-infinite
interval
supported
on the whole
real line
with support
whose type varies
Mixed
univariate
continuous-
discrete
Multivariate
(joint)
Directional
Degenerate
andsingular
Degenerate
Dirac delta function
Singular
Cantor
Families
Retrieved from "https://en.wikipedia.org/w/index.php?title=Negative_binomial_distribution&oldid=1335433204"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp