Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Pairwise independence

From Wikipedia, the free encyclopedia
Set of random variables of which any two are independent

Inprobability theory, apairwise independent collection ofrandom variables is a set of random variables any two of which areindependent.[1] Any collection ofmutually independent random variables is pairwise independent, but some pairwise independent collections are not mutually independent. Pairwise independent random variables with finitevariance areuncorrelated.

A pair of random variablesX andY areindependentif and only if the random vector (X,Y) withjointcumulative distribution function (CDF)FX,Y(x,y){\displaystyle F_{X,Y}(x,y)} satisfies

FX,Y(x,y)=FX(x)FY(y),{\displaystyle F_{X,Y}(x,y)=F_{X}(x)F_{Y}(y),}

or equivalently, their joint densityfX,Y(x,y){\displaystyle f_{X,Y}(x,y)} satisfies

fX,Y(x,y)=fX(x)fY(y).{\displaystyle f_{X,Y}(x,y)=f_{X}(x)f_{Y}(y).}

That is, the joint distribution is equal to the product of the marginal distributions.[2]

Unless it is not clear in context, in practice the modifier "mutual" is usually dropped so thatindependence meansmutual independence. A statement such as "X,Y,Z are independent random variables" means thatX,Y,Z are mutually independent.

Example

[edit]

Pairwise independence does not imply mutual independence, as shown by the following example attributed to S. Bernstein.[3]

SupposeX andY are two independent tosses of afair coin, where we designate 1 for heads and 0 for tails. Let the third random variableZ be equal to 1 if exactly one of those coin tosses resulted in "heads", and 0 otherwise (i.e.,Z=XY{\displaystyle Z=X\oplus Y}). Then jointly the triple (X,Y,Z) has the followingprobability distribution:

(X,Y,Z)={(0,0,0)with probability 1/4,(0,1,1)with probability 1/4,(1,0,1)with probability 1/4,(1,1,0)with probability 1/4.{\displaystyle (X,Y,Z)={\begin{cases}(0,0,0)&{\text{with probability}}\ 1/4,\\(0,1,1)&{\text{with probability}}\ 1/4,\\(1,0,1)&{\text{with probability}}\ 1/4,\\(1,1,0)&{\text{with probability}}\ 1/4.\end{cases}}}

Here themarginal probability distributions are identical:fX(0)=fY(0)=fZ(0)=1/2,{\displaystyle f_{X}(0)=f_{Y}(0)=f_{Z}(0)=1/2,} andfX(1)=fY(1)=fZ(1)=1/2.{\displaystyle f_{X}(1)=f_{Y}(1)=f_{Z}(1)=1/2.} Thebivariate distributions also agree:fX,Y=fX,Z=fY,Z,{\displaystyle f_{X,Y}=f_{X,Z}=f_{Y,Z},} wherefX,Y(0,0)=fX,Y(0,1)=fX,Y(1,0)=fX,Y(1,1)=1/4.{\displaystyle f_{X,Y}(0,0)=f_{X,Y}(0,1)=f_{X,Y}(1,0)=f_{X,Y}(1,1)=1/4.}

Since each of the pairwise joint distributions equals the product of their respective marginal distributions, the variables are pairwise independent:

  • X andY are independent, and
  • X andZ are independent, and
  • Y andZ are independent.

However,X,Y, andZ arenotmutually independent, sincefX,Y,Z(x,y,z)fX(x)fY(y)fZ(z),{\displaystyle f_{X,Y,Z}(x,y,z)\neq f_{X}(x)f_{Y}(y)f_{Z}(z),} the left side equalling for example 1/4 for (x,y,z) = (0, 0, 0) while the right side equals 1/8 for (x,y,z) = (0, 0, 0). In fact, any of{X,Y,Z}{\displaystyle \{X,Y,Z\}} is completely determined by the other two (any ofX,Y,Z is thesum (modulo 2) of the others). That is as far from independence as random variables can get.

Probability of the union of pairwise independent events

[edit]

Bounds on theprobability that the sum ofBernoullirandom variables is at least one, commonly known as theunion bound, are provided by theBoole–Fréchet[4][5] inequalities. While these bounds assume onlyunivariate information, several bounds with knowledge of generalbivariate probabilities, have been proposed too. Denote by{Ai,i{1,2,...,n}}{\displaystyle \{{A}_{i},i\in \{1,2,...,n\}\}} a set ofn{\displaystyle n}Bernoulli events withprobability of occurrenceP(Ai)=pi{\displaystyle \mathbb {P} (A_{i})=p_{i}} for eachi{\displaystyle i}. Suppose thebivariate probabilities are given byP(AiAj)=pij{\displaystyle \mathbb {P} (A_{i}\cap A_{j})=p_{ij}} for every pair of indices(i,j){\displaystyle (i,j)}. Kounias[6] derived the followingupper bound:P(iAi)i=1npimaxj{1,2,..,n}ijpij,{\displaystyle \mathbb {P} (\displaystyle {\cup }_{i}A_{i})\leq \displaystyle \sum _{i=1}^{n}p_{i}-{\underset {j\in \{1,2,..,n\}}{\max }}\sum _{i\neq j}p_{ij},}which subtracts the maximum weight of astarspanning tree on acomplete graph withn{\displaystyle n} nodes (where the edge weights are given bypij{\displaystyle p_{ij}}) from the sum of themarginal probabilitiesipi{\displaystyle \sum _{i}p_{i}}.
Hunter-Worsley[7][8] tightened thisupper bound by optimizing overτT{\displaystyle \tau \in T} as follows:P(iAi)i=1npimaxτT(i,j)τpij,{\displaystyle \mathbb {P} (\displaystyle {\cup }_{i}A_{i})\leq \displaystyle \sum _{i=1}^{n}p_{i}-{\underset {\tau \in T}{\max }}\sum _{(i,j)\in \tau }p_{ij},}whereT{\displaystyle T} is the set of allspanning trees on the graph. These bounds are not thetightest possible with generalbivariatespij{\displaystyle p_{ij}} even whenfeasibility is guaranteed as shown in Boros et.al.[9] However, when the variables arepairwise independent (pij=pipj{\displaystyle p_{ij}=p_{i}p_{j}}), Ramachandra—Natarajan[10] showed that the Kounias-Hunter-Worsley[6][7][8] bound istight by proving that the maximum probability of the union of events admits aclosed-form expression given as:

maxP(iAi)=min(i=1npipn(i=1n1pi),1){\displaystyle \max \mathbb {P} (\displaystyle {\cup }_{i}A_{i})=\displaystyle \min \left(\sum _{i=1}^{n}p_{i}-p_{n}\left(\sum _{i=1}^{n-1}p_{i}\right),1\right)}1

where theprobabilities are sorted in increasing order as0p1p2pn1{\displaystyle 0\leq p_{1}\leq p_{2}\leq \ldots \leq p_{n}\leq 1}. Thetight bound inEq. 1 depends only on the sum of the smallestn1{\displaystyle n-1}probabilitiesi=1n1pi{\displaystyle \sum _{i=1}^{n-1}p_{i}} and the largest probabilitypn{\displaystyle p_{n}}. Thus, whileordering of theprobabilities plays a role in the derivation of the bound, theordering among the smallestn1{\displaystyle n-1}probabilities{p1,p2,,pn1}{\displaystyle \{p_{1},p_{2},\dots ,p_{n-1}\}} is inconsequential since only their sum is used.

Comparison with theBoole–Fréchetunion bound

[edit]

It is useful to compare the smallest bounds on the probability of the union with arbitrarydependence andpairwise independence respectively. ThetightestBoole–Fréchetupperunion bound (assuming onlyunivariate information) is given as:

maxP(iAi)=min(i=1npi,1){\displaystyle \displaystyle \max \mathbb {P} (\displaystyle {\cup }_{i}A_{i})=\displaystyle \min \left(\sum _{i=1}^{n}p_{i},1\right)}2

As shown in Ramachandra-Natarajan,[10] it can be easily verified that the ratio of the twotight bounds inEq. 2 andEq. 1 isupper bounded by4/3{\displaystyle 4/3} where the maximum value of4/3{\displaystyle 4/3} is attained wheni=1n1pi=1/2,pn=1/2{\displaystyle \sum _{i=1}^{n-1}p_{i}=1/2\,,\quad p_{n}=1/2}where theprobabilities are sorted in increasing order as0p1p2pn1{\displaystyle 0\leq p_{1}\leq p_{2}\leq \dots \leq p_{n}\leq 1}. In other words, in the best-case scenario, the pairwise independence bound inEq. 1 provides an improvement of25%{\displaystyle 25\%} over theunivariate bound inEq. 2.

Generalization

[edit]

More generally, we can talk aboutk-wise independence, for anyk ≥ 2. The idea is similar: a set ofrandom variables isk-wise independent if every subset of sizek of those variables is independent.k-wise independence has been used in theoretical computer science, where it was used to prove a theorem about the problemMAXEkSAT.

k-wise independence is used in the proof thatk-independent hashing functions are secure unforgeablemessage authentication codes.

See also

[edit]

References

[edit]
  1. ^Gut, A. (2005)Probability: a Graduate Course, Springer-Verlag.ISBN 0-387-27332-8. pp. 71–72.
  2. ^Hogg, R. V., McKean, J. W., Craig, A. T. (2005).Introduction to Mathematical Statistics (6 ed.). Upper Saddle River, NJ: Pearson Prentice Hall.ISBN 0-13-008507-3.{{cite book}}: CS1 maint: multiple names: authors list (link) Definition 2.5.1, page 109.
  3. ^Hogg, R. V., McKean, J. W., Craig, A. T. (2005).Introduction to Mathematical Statistics (6 ed.). Upper Saddle River, NJ: Pearson Prentice Hall.ISBN 0-13-008507-3.{{cite book}}: CS1 maint: multiple names: authors list (link) Remark 2.6.1, p. 120.
  4. ^Boole, G. (1854).An Investigation of the Laws of Thought, On Which Are Founded the Mathematical Theories of Logic and Probability. Walton and Maberly, London. See Boole's "major" and "minor" limits of a conjunction on page 299.
  5. ^Fréchet, M. (1935). Généralisations du théorème des probabilités totales.Fundamenta Mathematicae25: 379–387.
  6. ^abE. G. Kounias (1968)."Bounds for the probability of a union, with applications".The Annals of Mathematical Statistics.39 (6):2154–2158.doi:10.1214/aoms/1177698049.
  7. ^abD. Hunter (1976). "An upper bound for the probability of a union".Journal of Applied Probability.13 (3):597–603.doi:10.2307/3212481.JSTOR 3212481.
  8. ^abK. J. Worsley (1982). "An improved Bonferroni inequality and applications".Biometrika.69 (2):297–302.doi:10.1093/biomet/69.2.297.
  9. ^Boros, Endre; Scozzari, Andrea; Tardella, Fabio; Veneziani, Pierangela (2014). "Polynomially computable bounds for the probability of the union of events".Mathematics of Operations Research.39 (4):1311–1329.doi:10.1287/moor.2014.0657.
  10. ^abRamachandra, Arjun Kodagehalli; Natarajan, Karthik (2023). "Tight Probability Bounds with Pairwise Independence".SIAM Journal on Discrete Mathematics.37 (2):516–555.arXiv:2006.00516.doi:10.1137/21M140829.
Retrieved from "https://en.wikipedia.org/w/index.php?title=Pairwise_independence&oldid=1333488831"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp