Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Random variable

From Wikipedia, the free encyclopedia
Variable representing a random phenomenon
Part of a series onstatistics
Probability theory

Arandom variable (also calledrandom quantity,aleatory variable, orstochastic variable) is a mathematical formalization of a quantity or object which depends onrandom events.[1] The term 'random variable' in its mathematical definition refers to neither randomness nor variability[2] but instead is a mathematicalfunction in which

This graph shows how random variable is a function from all possible outcomes to real values. It also shows how random variable is used for defining probability mass functions.

Informally, randomness typically represents some fundamental element of chance, such as in the roll of adie; it may also represent uncertainty, such asmeasurement error.[1] However, theinterpretation of probability is philosophically complicated, and even in specific cases is not always straightforward. The purely mathematical analysis of random variables is independent of such interpretational difficulties, and can be based upon a rigorousaxiomatic setup.

In the formal mathematical language ofmeasure theory, a random variable is defined as ameasurable function from aprobability measure space (called thesample space) to ameasurable space. This allows consideration of thepushforward measure, which is called thedistribution of the random variable; the distribution is thus aprobability measure on the set of all possible values of the random variable. It is possible for two random variables to have identical distributions but to differ in significant ways; for instance, they may beindependent.

It is common to consider the special cases ofdiscrete random variables andabsolutely continuous random variables, corresponding to whether a random variable is valued in a countable subset or in an interval ofreal numbers. There are other important possibilities, especially in the theory ofstochastic processes, wherein it is natural to considerrandom sequences orrandom functions. Sometimes arandom variable is taken to be automatically valued in the real numbers, with more general random quantities instead being calledrandom elements.

According toGeorge Mackey,Pafnuty Chebyshev was the first person "to think systematically in terms of random variables".[3]

Definition

[edit]

Arandom variableX{\displaystyle X} is ameasurable functionX:ΩE{\displaystyle X\colon \Omega \to E} from a sample spaceΩ{\displaystyle \Omega } as a set of possibleoutcomes to ameasurable spaceE{\displaystyle E}. The technical axiomatic definition requires the sample spaceΩ{\displaystyle \Omega } to belong to aprobability triple(Ω,F,P){\displaystyle (\Omega ,{\mathcal {F}},\operatorname {P} )} (see themeasure-theoretic definition). A random variable is often denoted by capitalRoman letters such asX,Y,Z,T{\displaystyle X,Y,Z,T}.[4]

The probability thatX{\displaystyle X} takes on a value in a measurable setSE{\displaystyle S\subseteq E} is written as

P(XS)=P({ωΩX(ω)S}){\displaystyle \operatorname {P} (X\in S)=\operatorname {P} (\{\omega \in \Omega \mid X(\omega )\in S\})}.

Standard case

[edit]

In many cases,X{\displaystyle X} isreal-valued, i.e.E=R{\displaystyle E=\mathbb {R} }. In some contexts, the termrandom element (seeextensions) is used to denote a random variable not of this form.

When theimage (or range) ofX{\displaystyle X} is finitely or infinitelycountable, the random variable is called adiscrete random variable[5]: 399  and its distribution is adiscrete probability distribution, i.e. can be described by aprobability mass function that assigns a probability to each value in the image ofX{\displaystyle X}. If the image is uncountably infinite (usually aninterval) thenX{\displaystyle X} is called acontinuous random variable.[6][7] In the special case that it isabsolutely continuous, its distribution can be described by aprobability density function, which assigns probabilities to intervals; in particular, each individual point must necessarily have probability zero for an absolutely continuous random variable. Not all continuous random variables are absolutely continuous.[8]

Any random variable can be described by itscumulative distribution function, which describes the probability that the random variable will be less than or equal to a certain value.

Extensions

[edit]

The term "random variable" in statistics is traditionally limited to thereal-valued case (E=R{\displaystyle E=\mathbb {R} }). In this case, the structure of the real numbers makes it possible to define quantities such as theexpected value andvariance of a random variable, itscumulative distribution function, and themoments of its distribution.

However, the definition above is valid for anymeasurable spaceE{\displaystyle E} of values. Thus one can consider random elements of other setsE{\displaystyle E}, such as randomBoolean values,categorical values,complex numbers,vectors,matrices,sequences,trees,sets,shapes,manifolds, andfunctions. One may then specifically refer to arandom variable oftypeE{\displaystyle E}, or anE{\displaystyle E}-valued random variable.

This more general concept of arandom element is particularly useful in disciplines such asgraph theory,machine learning,natural language processing, and other fields indiscrete mathematics andcomputer science, where one is often interested in modeling the random variation of non-numericaldata structures. In some cases, it is nonetheless convenient to represent each element ofE{\displaystyle E}, using one or more real numbers. In this case, a random element may optionally be represented as avector of real-valued random variables (all defined on the same underlying probability spaceΩ{\displaystyle \Omega }, which allows the different random variables tocovary). For example:

Distribution functions

[edit]

If a random variableX:ΩR{\displaystyle X\colon \Omega \to \mathbb {R} } defined on the probability space(Ω,F,P){\displaystyle (\Omega ,{\mathcal {F}},\operatorname {P} )} is given, we can ask questions like "How likely is it that the value ofX{\displaystyle X} is equal to 2?". This is the same as the probability of the event{ω:X(ω)=2}{\displaystyle \{\omega :X(\omega )=2\}\,\!} which is often written asP(X=2){\displaystyle P(X=2)\,\!} orpX(2){\displaystyle p_{X}(2)} for short.

Recording all these probabilities of outputs of a random variableX{\displaystyle X} yields theprobability distribution ofX{\displaystyle X}. The probability distribution "forgets" about the particular probability space used to defineX{\displaystyle X} and only records the probabilities of various output values ofX{\displaystyle X}. Such a probability distribution, ifX{\displaystyle X} is real-valued, can always be captured by itscumulative distribution function

FX(x)=P(Xx){\displaystyle F_{X}(x)=\operatorname {P} (X\leq x)}

and sometimes also using aprobability density function,fX{\displaystyle f_{X}}. Inmeasure-theoretic terms, we use the random variableX{\displaystyle X} to "push-forward" the measureP{\displaystyle P} onΩ{\displaystyle \Omega } to a measurepX{\displaystyle p_{X}} onR{\displaystyle \mathbb {R} }. The measurepX{\displaystyle p_{X}} is called the "(probability) distribution ofX{\displaystyle X}" or the "law ofX{\displaystyle X}".[9] The densityfX=dpX/dμ{\displaystyle f_{X}=dp_{X}/d\mu }, theRadon–Nikodym derivative ofpX{\displaystyle p_{X}} with respect to some reference measureμ{\displaystyle \mu } onR{\displaystyle \mathbb {R} } (often, this reference measure is theLebesgue measure in the case of continuous random variables, or thecounting measure in the case of discrete random variables).The underlying probability spaceΩ{\displaystyle \Omega } is a technical device used to guarantee the existence of random variables, sometimes to construct them, and to define notions such ascorrelation and dependence orindependence based on ajoint distribution of two or more random variables on the same probability space. In practice, one often disposes of the spaceΩ{\displaystyle \Omega } altogether and just puts a measure onR{\displaystyle \mathbb {R} } that assigns measure 1 to the whole real line, i.e., one works with probability distributions instead of random variables. See the article onquantile functions for fuller development.

Examples

[edit]

Discrete random variable

[edit]

Consider an experiment where a person is chosen at random. An example of a random variable may be the person's height. Mathematically, the random variable is interpreted as a function which maps the person to their height. Associated with the random variable is a probability distribution that allows the computation of the probability that the height is in any subset of possible values, such as the probability that the height is between 180 and 190 cm, or the probability that the height is either less than 150 or more than 200 cm.

Another random variable may be the person's number of children; this is a discrete random variable with non-negative integer values. It allows the computation of probabilities for individual integer values – the probability mass function (PMF) – or for sets of values, including infinite sets. For example, the event of interest may be "an even number of children". For both finite and infinite event sets, their probabilities can be found by adding up the PMFs of the elements; that is, the probability of an even number of children is the infinite sumPMF(0)+PMF(2)+PMF(4)+{\displaystyle \operatorname {PMF} (0)+\operatorname {PMF} (2)+\operatorname {PMF} (4)+\cdots }.

In examples such as these, thesample space is often suppressed, since it is mathematically hard to describe, and the possible values of the random variables are then treated as a sample space. But when two random variables are measured on the same sample space of outcomes, such as the height and number of children being computed on the same random persons, it is easier to track their relationship if it is acknowledged that both height and number of children come from the same random person, for example so that questions of whether such random variables are correlated or not can be posed.

If{an},{bn}{\textstyle \{a_{n}\},\{b_{n}\}} are countable sets of real numbers,bn>0{\textstyle b_{n}>0} andnbn=1{\textstyle \sum _{n}b_{n}=1}, thenF=nbnδan(x){\textstyle F=\sum _{n}b_{n}\delta _{a_{n}}(x)} is a discrete distribution function. Hereδt(x)=0{\displaystyle \delta _{t}(x)=0} forx<t{\displaystyle x<t},δt(x)=1{\displaystyle \delta _{t}(x)=1} forxt{\displaystyle x\geq t}. Taking for instance an enumeration of all rational numbers as{an}{\displaystyle \{a_{n}\}} , one gets a discrete function that is not necessarily a step function (piecewise constant).

Coin toss

[edit]

The possible outcomes for one coin toss can be described by the sample spaceΩ={heads,tails}{\displaystyle \Omega =\{{\text{heads}},{\text{tails}}\}}. We can introduce a real-valued random variableY{\displaystyle Y} that models a $1 payoff for a successful bet on heads as follows:Y(ω)={1,if ω=heads,0,if ω=tails.{\displaystyle Y(\omega )={\begin{cases}1,&{\text{if }}\omega ={\text{heads}},\\[6pt]0,&{\text{if }}\omega ={\text{tails}}.\end{cases}}}

If the coin is afair coin,Y has aprobability mass functionfY{\displaystyle f_{Y}} given by:fY(y)={12,if y=1,12,if y=0,{\displaystyle f_{Y}(y)={\begin{cases}{\tfrac {1}{2}},&{\text{if }}y=1,\\[6pt]{\tfrac {1}{2}},&{\text{if }}y=0,\end{cases}}}

Dice roll

[edit]
If the sample space is the set of possible numbers rolled on two dice, and the random variable of interest is the sumS of the numbers on the two dice, thenS is a discrete random variable whose distribution is described by theprobability mass function plotted as the height of picture columns here.

A random variable can also be used to describe the process of rolling dice and the possible outcomes. The most obvious representation for the two-dice case is to take the set of pairs of numbersn1 andn2 from {1, 2, 3, 4, 5, 6} (representing the numbers on the two dice) as the sample space. The total number rolled (the sum of the numbers in each pair) is then a random variableX given by the function that maps the pair to the sum:X((n1,n2))=n1+n2{\displaystyle X((n_{1},n_{2}))=n_{1}+n_{2}}and (if the dice arefair) has a probability mass functionfX given by:fX(S)=min(S1,13S)36, for S{2,3,4,5,6,7,8,9,10,11,12}{\displaystyle f_{X}(S)={\frac {\min(S-1,13-S)}{36}},{\text{ for }}S\in \{2,3,4,5,6,7,8,9,10,11,12\}}

Continuous random variable

[edit]

Formally, a continuous random variable is a random variable whosecumulative distribution function iscontinuous everywhere.[10] There are no "gaps", which would correspond to numbers which have a finite probability ofoccurring. Instead, continuous random variablesalmost never take an exact prescribed valuec (formally,cR:Pr(X=c)=0{\textstyle \forall c\in \mathbb {R} :\;\Pr(X=c)=0}) but there is a positive probability that its value will lie in particularintervals which can bearbitrarily small. Continuous random variables usually admitprobability density functions (PDF), which characterize their CDF andprobability measures; such distributions are also calledabsolutely continuous; but some continuous distributions aresingular, or mixes of an absolutely continuous part and a singular part.

An example of a continuous random variable would be one based on a spinner that can choose a horizontal direction. Then the values taken by the random variable are directions. We could represent these directions by North, West, East, South, Southeast, etc. However, it is commonly more convenient to map the sample space to a random variable which takes values which are real numbers. This can be done, for example, by mapping a direction to a bearing in degrees clockwise from North. The random variable then takes values which are real numbers from the interval [0, 360), with all parts of the range being "equally likely". In this case,X = the angle spun. Any real number has probability zero of being selected, but a positive probability can be assigned to anyrange of values. For example, the probability of choosing a number in [0, 180] is12. Instead of speaking of a probability mass function, we say that the probabilitydensity ofX is 1/360. The probability of a subset of [0, 360) can be calculated by multiplying the measure of the set by 1/360. In general, the probability of a set for a given continuous random variable can be calculated by integrating the density over the given set.

More formally, given anyintervalI=[a,b]={xR:axb}{\textstyle I=[a,b]=\{x\in \mathbb {R} :a\leq x\leq b\}}, a random variableXIU(I)=U[a,b]{\displaystyle X_{I}\sim \operatorname {U} (I)=\operatorname {U} [a,b]} is called a "continuous uniform random variable" (CURV) if the probability that it takes a value in asubinterval depends only on the length of the subinterval. This implies that the probability ofXI{\displaystyle X_{I}} falling in any subinterval[c,d][a,b]{\displaystyle [c,d]\subseteq [a,b]} isproportional to thelength of the subinterval, that is, ifacdb, one has

Pr(XI[c,d])=dcba{\displaystyle \Pr \left(X_{I}\in [c,d]\right)={\frac {d-c}{b-a}}}

where the last equality results from theunitarity axiom of probability. Theprobability density function of a CURVXU[a,b]{\displaystyle X\sim \operatorname {U} [a,b]} is given by theindicator function of its interval ofsupport normalized by the interval's length:fX(x)={1ba,axb0,otherwise.{\displaystyle f_{X}(x)={\begin{cases}\displaystyle {1 \over b-a},&a\leq x\leq b\\0,&{\text{otherwise}}.\end{cases}}}Of particular interest is the uniform distribution on theunit interval[0,1]{\displaystyle [0,1]}. Samples of any desiredprobability distributionD{\displaystyle \operatorname {D} } can be generated by calculating thequantile function ofD{\displaystyle \operatorname {D} } on arandomly-generated number distributed uniformly on the unit interval. This exploitsproperties of cumulative distribution functions, which are a unifying framework for all random variables.

Mixed type

[edit]

Amixed random variable is a random variable whosecumulative distribution function is neitherdiscrete noreverywhere-continuous.[10] It can be realized as a mixture of a discrete random variable and a continuous random variable; in which case theCDF will be the weighted average of the CDFs of the component variables.[10]

An example of a random variable of mixed type would be based on an experiment where a coin is flipped and the spinner is spun only if the result of the coin toss is heads. If the result is tails,X = −1; otherwiseX = the value of the spinner as in the preceding example. There is a probability of12 that this random variable will have the value −1. Other ranges of values would have half the probabilities of the last example.

Most generally, every probability distribution on the real line is a mixture of discrete part, singular part, and an absolutely continuous part; seeLebesgue's decomposition theorem § Refinement. The discrete part is concentrated on a countable set, but this set may be dense (like the set of all rational numbers).

Measure-theoretic definition

[edit]

The most formal,axiomatic definition of a random variable involvesmeasure theory. Continuous random variables are defined in terms ofsets of numbers, along with functions that map such sets to probabilities. Because of various difficulties (e.g. theBanach–Tarski paradox) that arise if such sets are insufficiently constrained, it is necessary to introduce what is termed asigma-algebra to constrain the possible sets over which probabilities can be defined. Normally, a particular such sigma-algebra is used, theBorel σ-algebra, which allows for probabilities to be defined over any sets that can be derived either directly from continuous intervals of numbers or by a finite orcountably infinite number ofunions and/orintersections of such intervals.[11]

The measure-theoretic definition is as follows.

Let(Ω,F,P){\displaystyle (\Omega ,{\mathcal {F}},P)} be aprobability space and(E,E){\displaystyle (E,{\mathcal {E}})} ameasurable space. Then an(E,E){\displaystyle (E,{\mathcal {E}})}-valued random variable is a measurable functionX:ΩE{\displaystyle X\colon \Omega \to E}, which means that, for every subsetBE{\displaystyle B\in {\mathcal {E}}}, itspreimage isF{\displaystyle {\mathcal {F}}}-measurable;X1(B)F{\displaystyle X^{-1}(B)\in {\mathcal {F}}}, whereX1(B)={ω:X(ω)B}{\displaystyle X^{-1}(B)=\{\omega :X(\omega )\in B\}}.[12] This definition enables us to measure any subsetBE{\displaystyle B\in {\mathcal {E}}} in the target space by looking at its preimage, which by assumption is measurable.

In more intuitive terms, a member ofΩ{\displaystyle \Omega } is a possible outcome, a member ofF{\displaystyle {\mathcal {F}}} is a measurable subset of possible outcomes, the functionP{\displaystyle P} gives the probability of each such measurable subset,E{\displaystyle E} represents the set of values that the random variable can take (such as the set of real numbers), and a member ofE{\displaystyle {\mathcal {E}}} is a "well-behaved" (measurable) subset ofE{\displaystyle E} (those for which the probability may be determined). The random variable is then a function from any outcome to a quantity, such that the outcomes leading to any useful subset of quantities for the random variable have a well-defined probability.

WhenE{\displaystyle E} is atopological space, then the most common choice for theσ-algebraE{\displaystyle {\mathcal {E}}} is theBorel σ-algebraB(E){\displaystyle {\mathcal {B}}(E)}, which is the σ-algebra generated by the collection of all open sets inE{\displaystyle E}. In such case the(E,E){\displaystyle (E,{\mathcal {E}})}-valued random variable is called anE{\displaystyle E}-valued random variable. Moreover, when the spaceE{\displaystyle E} is the real lineR{\displaystyle \mathbb {R} }, then such a real-valued random variable is called simply arandom variable.

Real-valued random variables

[edit]

In this case the observation space is the set of real numbers. Recall,(Ω,F,P){\displaystyle (\Omega ,{\mathcal {F}},P)} is the probability space. For a real observation space, the functionX:ΩR{\displaystyle X\colon \Omega \rightarrow \mathbb {R} } is a real-valued random variable if

{ω:X(ω)r}FrR.{\displaystyle \{\omega :X(\omega )\leq r\}\in {\mathcal {F}}\qquad \forall r\in \mathbb {R} .}

This definition is a special case of the above because the set{(,r]:rR}{\displaystyle \{(-\infty ,r]:r\in \mathbb {R} \}} generates the Borel σ-algebra on the set of real numbers, and it suffices to check measurability on any generating set. Here we can prove measurability on this generating set by using the fact that{ω:X(ω)r}=X1((,r]){\displaystyle \{\omega :X(\omega )\leq r\}=X^{-1}((-\infty ,r])}.

Moments

[edit]

The probability distribution of a random variable is often characterised by a small number of parameters, which also have a practical interpretation. For example, it is often enough to know what its "average value" is. This is captured by the mathematical concept ofexpected value of a random variable, denotedE[X]{\displaystyle \operatorname {E} [X]}, and also called thefirstmoment. In general,E[f(X)]{\displaystyle \operatorname {E} [f(X)]} is not equal tof(E[X]){\displaystyle f(\operatorname {E} [X])}. Once the "average value" is known, one could then ask how far from this average value the values ofX{\displaystyle X} typically are, a question that is answered by thevariance andstandard deviation of a random variable.E[X]{\displaystyle \operatorname {E} [X]} can be viewed intuitively as an average obtained from an infinite population, the members of which are particular evaluations ofX{\displaystyle X}.

Mathematically, this is known as the (generalised)problem of moments: for a given class of random variablesX{\displaystyle X}, find a collection{fi}{\displaystyle \{f_{i}\}} of functions such that the expectation valuesE[fi(X)]{\displaystyle \operatorname {E} [f_{i}(X)]} fully characterise thedistribution of the random variableX{\displaystyle X}.

Moments can only be defined for real-valued functions of random variables (or complex-valued, etc.). If the random variable is itself real-valued, then moments of the variable itself can be taken, which are equivalent to moments of the identity functionf(X)=X{\displaystyle f(X)=X} of the random variable. However, even for non-real-valued random variables, moments can be taken of real-valued functions of those variables. For example, for acategorical random variableX that can take on thenominal values "red", "blue" or "green", the real-valued function[X=green]{\displaystyle [X={\text{green}}]} can be constructed; this uses theIverson bracket, and has the value 1 ifX{\displaystyle X} has the value "green", 0 otherwise. Then, theexpected value and other moments of this function can be determined.

Functions of random variables

[edit]

A new random variableY can be defined byapplying a realBorel measurable functiong:RR{\displaystyle g\colon \mathbb {R} \rightarrow \mathbb {R} } to the outcomes of areal-valued random variableX{\displaystyle X}. That is,Y=g(X){\displaystyle Y=g(X)}. Thecumulative distribution function ofY{\displaystyle Y} is then

FY(y)=P(g(X)y).{\displaystyle F_{Y}(y)=\operatorname {P} (g(X)\leq y).}

If functiong{\displaystyle g} is invertible (i.e.,h=g1{\displaystyle h=g^{-1}} exists, whereh{\displaystyle h} isg{\displaystyle g}'sinverse function) and is eitherincreasing or decreasing, then the previous relation can be extended to obtain

FY(y)=P(g(X)y)={P(Xh(y))=FX(h(y)),if h=g1 increasing,P(Xh(y))=1FX(h(y)),if h=g1 decreasing.{\displaystyle F_{Y}(y)=\operatorname {P} (g(X)\leq y)={\begin{cases}\operatorname {P} (X\leq h(y))=F_{X}(h(y)),&{\text{if }}h=g^{-1}{\text{ increasing}},\\\\\operatorname {P} (X\geq h(y))=1-F_{X}(h(y)),&{\text{if }}h=g^{-1}{\text{ decreasing}}.\end{cases}}}

With the same hypotheses of invertibility ofg{\displaystyle g}, assuming alsodifferentiability, the relation between theprobability density functions can be found by differentiating both sides of the above expression with respect toy{\displaystyle y}, in order to obtain[10]

fY(y)=fX(h(y))|dh(y)dy|.{\displaystyle f_{Y}(y)=f_{X}{\bigl (}h(y){\bigr )}\left|{\frac {dh(y)}{dy}}\right|.}

If there is no invertibility ofg{\displaystyle g} but eachy{\displaystyle y} admits at most a countable number of roots (i.e., a finite, or countably infinite, number ofxi{\displaystyle x_{i}} such thaty=g(xi){\displaystyle y=g(x_{i})}) then the previous relation between theprobability density functions can be generalized with

fY(y)=ifX(gi1(y))|dgi1(y)dy|{\displaystyle f_{Y}(y)=\sum _{i}f_{X}(g_{i}^{-1}(y))\left|{\frac {dg_{i}^{-1}(y)}{dy}}\right|}

wherexi=gi1(y){\displaystyle x_{i}=g_{i}^{-1}(y)}, according to theinverse function theorem. The formulas for densities do not demandg{\displaystyle g} to be increasing.

In the measure-theoretic,axiomatic approach to probability, if a random variableX{\displaystyle X} onΩ{\displaystyle \Omega } and aBorel measurable functiong:RR{\displaystyle g\colon \mathbb {R} \rightarrow \mathbb {R} }, thenY=g(X){\displaystyle Y=g(X)} is also a random variable onΩ{\displaystyle \Omega }, since the composition of measurable functionsis also measurable. (However, this is not necessarily true ifg{\displaystyle g} isLebesgue measurable.[citation needed]) The same procedure that allowed one to go from a probability space(Ω,P){\displaystyle (\Omega ,P)} to(R,dFX){\displaystyle (\mathbb {R} ,dF_{X})} can be used to obtain the distribution ofY{\displaystyle Y}.

Example 1

[edit]

LetX{\displaystyle X} be a real-valued,continuous random variable and letY=X2{\displaystyle Y=X^{2}}.

FY(y)=P(X2y).{\displaystyle F_{Y}(y)=\operatorname {P} (X^{2}\leq y).}

Ify<0{\displaystyle y<0}, thenP(X2y)=0{\displaystyle P(X^{2}\leq y)=0}, so

FY(y)=0ify<0.{\displaystyle F_{Y}(y)=0\qquad {\hbox{if}}\quad y<0.}

Ify0{\displaystyle y\geq 0}, then

P(X2y)=P(|X|y)=P(yXy),{\displaystyle \operatorname {P} (X^{2}\leq y)=\operatorname {P} (|X|\leq {\sqrt {y}})=\operatorname {P} (-{\sqrt {y}}\leq X\leq {\sqrt {y}}),}

so

FY(y)=FX(y)FX(y)ify0.{\displaystyle F_{Y}(y)=F_{X}({\sqrt {y}})-F_{X}(-{\sqrt {y}})\qquad {\hbox{if}}\quad y\geq 0.}

Example 2

[edit]

SupposeX{\displaystyle X} is a random variable with a cumulative distribution

FX(x)=P(Xx)=1(1+ex)θ{\displaystyle F_{X}(x)=P(X\leq x)={\frac {1}{(1+e^{-x})^{\theta }}}}

whereθ>0{\displaystyle \theta >0} is a fixed parameter. Consider the random variableY=log(1+eX).{\displaystyle Y=\mathrm {log} (1+e^{-X}).} Then,

FY(y)=P(Yy)=P(log(1+eX)y)=P(Xlog(ey1)).{\displaystyle F_{Y}(y)=P(Y\leq y)=P(\mathrm {log} (1+e^{-X})\leq y)=P(X\geq -\mathrm {log} (e^{y}-1)).\,}

The last expression can be calculated in terms of the cumulative distribution ofX,{\displaystyle X,} so

FY(y)=1FX(log(ey1))=11(1+elog(ey1))θ=11(1+ey1)θ=1eyθ.{\displaystyle {\begin{aligned}F_{Y}(y)&=1-F_{X}(-\log(e^{y}-1))\\[5pt]&=1-{\frac {1}{(1+e^{\log(e^{y}-1)})^{\theta }}}\\[5pt]&=1-{\frac {1}{(1+e^{y}-1)^{\theta }}}\\[5pt]&=1-e^{-y\theta }.\end{aligned}}}

which is thecumulative distribution function (CDF) of anexponential distribution.

Example 3

[edit]

SupposeX{\displaystyle X} is a random variable with astandard normal distribution, whose density is

fX(x)=12πex2/2.{\displaystyle f_{X}(x)={\frac {1}{\sqrt {2\pi }}}e^{-x^{2}/2}.}

Consider the random variableY=X2.{\displaystyle Y=X^{2}.} We can find the density using the above formula for a change of variables:

fY(y)=ifX(gi1(y))|dgi1(y)dy|.{\displaystyle f_{Y}(y)=\sum _{i}f_{X}(g_{i}^{-1}(y))\left|{\frac {dg_{i}^{-1}(y)}{dy}}\right|.}

In this case the change is notmonotonic, because every value ofY{\displaystyle Y} has two corresponding values ofX{\displaystyle X} (one positive and negative). However, because of symmetry, both halves will transform identically, i.e.,

fY(y)=2fX(g1(y))|dg1(y)dy|.{\displaystyle f_{Y}(y)=2f_{X}(g^{-1}(y))\left|{\frac {dg^{-1}(y)}{dy}}\right|.}

The inverse transformation is

x=g1(y)=y{\displaystyle x=g^{-1}(y)={\sqrt {y}}}

and its derivative is

dg1(y)dy=12y.{\displaystyle {\frac {dg^{-1}(y)}{dy}}={\frac {1}{2{\sqrt {y}}}}.}

Then,

fY(y)=212πey/212y=12πyey/2.{\displaystyle f_{Y}(y)=2{\frac {1}{\sqrt {2\pi }}}e^{-y/2}{\frac {1}{2{\sqrt {y}}}}={\frac {1}{\sqrt {2\pi y}}}e^{-y/2}.}

This is achi-squared distribution with onedegree of freedom.

Example 4

[edit]

SupposeX{\displaystyle X} is a random variable with anormal distribution, whose density is

fX(x)=12πσ2e(xμ)2/(2σ2).{\displaystyle f_{X}(x)={\frac {1}{\sqrt {2\pi \sigma ^{2}}}}e^{-(x-\mu )^{2}/(2\sigma ^{2})}.}

Consider the random variableY=X2.{\displaystyle Y=X^{2}.} We can find the density using the above formula for a change of variables:

fY(y)=ifX(gi1(y))|dgi1(y)dy|.{\displaystyle f_{Y}(y)=\sum _{i}f_{X}(g_{i}^{-1}(y))\left|{\frac {dg_{i}^{-1}(y)}{dy}}\right|.}

In this case the change is notmonotonic, because every value ofY{\displaystyle Y} has two corresponding values ofX{\displaystyle X} (one positive and negative). Differently from the previous example, in this case however, there is no symmetry and we have to compute the two distinct terms:

fY(y)=fX(g11(y))|dg11(y)dy|+fX(g21(y))|dg21(y)dy|.{\displaystyle f_{Y}(y)=f_{X}(g_{1}^{-1}(y))\left|{\frac {dg_{1}^{-1}(y)}{dy}}\right|+f_{X}(g_{2}^{-1}(y))\left|{\frac {dg_{2}^{-1}(y)}{dy}}\right|.}

The inverse transformation is

x=g1,21(y)=±y{\displaystyle x=g_{1,2}^{-1}(y)=\pm {\sqrt {y}}}

and its derivative is

dg1,21(y)dy=±12y.{\displaystyle {\frac {dg_{1,2}^{-1}(y)}{dy}}=\pm {\frac {1}{2{\sqrt {y}}}}.}

Then,

fY(y)=12πσ212y(e(yμ)2/(2σ2)+e(yμ)2/(2σ2)).{\displaystyle f_{Y}(y)={\frac {1}{\sqrt {2\pi \sigma ^{2}}}}{\frac {1}{2{\sqrt {y}}}}(e^{-({\sqrt {y}}-\mu )^{2}/(2\sigma ^{2})}+e^{-(-{\sqrt {y}}-\mu )^{2}/(2\sigma ^{2})}).}

This is anoncentral chi-squared distribution with onedegree of freedom.

Some properties

[edit]
  • The probability distribution of the sum of two independent random variables is theconvolution of each of their distributions.
  • Probability distributions are not avector space—they are not closed underlinear combinations, as these do not preserve non-negativity or total integral 1—but they are closed underconvex combination, thus forming aconvex subset of the space of functions (or measures).

Equivalence of random variables

[edit]

There are several different senses in which random variables can be considered to be equivalent. Two random variables can be equal, equal almost surely, or equal in distribution.

In increasing order of strength, the precise definition of these notions of equivalence is given below.

Equality in distribution

[edit]

If the sample space is a subset of the real line, random variablesX andY areequal in distribution (denotedX=dY{\displaystyle X{\stackrel {d}{=}}Y}) if they have the same distribution functions:

P(Xx)=P(Yx)for all x.{\displaystyle \operatorname {P} (X\leq x)=\operatorname {P} (Y\leq x)\quad {\text{for all }}x.}

To be equal in distribution, random variables need not be defined on the same probability space. Two random variables having equalmoment generating functions have the same distribution. This provides, for example, a useful method of checking equality of certain functions ofindependent, identically distributed (IID) random variables. However, the moment generating function exists only for distributions that have a definedLaplace transform.

Almost sure equality

[edit]

Two random variablesX andY areequalalmost surely (denotedX=a.s.Y{\displaystyle X\;{\stackrel {\text{a.s.}}{=}}\;Y}) if, and only if, the probability that they are different iszero:

P(XY)=0.{\displaystyle \operatorname {P} (X\neq Y)=0.}

For all practical purposes in probability theory, this notion of equivalence is as strong as actual equality. It is associated to the following distance:

d(X,Y)=esssupω|X(ω)Y(ω)|,{\displaystyle d_{\infty }(X,Y)=\operatorname {ess} \sup _{\omega }|X(\omega )-Y(\omega )|,}

where "ess sup" represents theessential supremum in the sense ofmeasure theory.

Equality

[edit]

Finally, the two random variablesX andY areequal if they are equal as functions on their measurable space:

X(ω)=Y(ω)for all ω.{\displaystyle X(\omega )=Y(\omega )\qquad {\hbox{for all }}\omega .}

This notion is typically the least useful in probability theory because in practice and in theory, the underlyingmeasure space of theexperiment is rarely explicitly characterized or even characterizable.

Practical difference between notions of equivalence

[edit]

Since we rarely explicitly construct the probability space underlying a random variable, the difference between these notions of equivalence is somewhat subtle. Essentially, two random variables consideredin isolation are "practically equivalent" if they are equal in distribution -- but once we relate them toother random variables defined on the same probability space, then they only remain "practically equivalent" if they are equal almost surely.

For example, consider the real random variablesA,B,C, andD all defined on the same probability space. Suppose thatA andB are equal almost surely (A=a.s.B{\displaystyle A\;{\stackrel {\text{a.s.}}{=}}\;B}), butA andC are only equal in distribution (A=dC{\displaystyle A{\stackrel {d}{=}}C}). ThenA+D=a.s.B+D{\displaystyle A+D\;{\stackrel {\text{a.s.}}{=}}\;B+D}, but in generalA+DC+D{\displaystyle A+D\;\neq \;C+D} (not even in distribution). Similarly, we have that the expectation valuesE(AD)=E(BD){\displaystyle \mathbb {E} (AD)=\mathbb {E} (BD)}, but in generalE(AD)E(CD){\displaystyle \mathbb {E} (AD)\neq \mathbb {E} (CD)}. Therefore, two random variables that are equal in distribution (but not equal almost surely) can have differentcovariances with a third random variable.

Convergence

[edit]
Main article:Convergence of random variables

A significant theme in mathematical statistics consists of obtaining convergence results for certainsequences of random variables; for instance thelaw of large numbers and thecentral limit theorem.

There are various senses in which a sequenceXn{\displaystyle X_{n}} of random variables can converge to a random variableX{\displaystyle X}. These are explained in the article onconvergence of random variables.

See also

[edit]

References

[edit]

Inline citations

[edit]
  1. ^abBlitzstein, Joe; Hwang, Jessica (2014).Introduction to Probability. CRC Press.ISBN 9781466575592.
  2. ^Deisenroth, Marc Peter (2020).Mathematics for machine learning. A. Aldo Faisal, Cheng Soon Ong. Cambridge, United Kingdom: Cambridge University Press.ISBN 978-1-108-47004-9.OCLC 1104219401.
  3. ^George Mackey (July 1980). "Harmonic analysis as the exploitation of symmetry – a historical survey".Bulletin of the American Mathematical Society. New Series.3 (1).
  4. ^"Random Variables".www.mathsisfun.com. Retrieved2020-08-21.
  5. ^Yates, Daniel S.; Moore, David S; Starnes, Daren S. (2003).The Practice of Statistics (2nd ed.). New York:Freeman.ISBN 978-0-7167-4773-4. Archived fromthe original on 2005-02-09.
  6. ^"Random Variables".www.stat.yale.edu. Retrieved2020-08-21.
  7. ^Dekking, Frederik Michel; Kraaikamp, Cornelis; Lopuhaä, Hendrik Paul; Meester, Ludolf Erwin (2005)."A Modern Introduction to Probability and Statistics".Springer Texts in Statistics.doi:10.1007/1-84628-168-7.ISBN 978-1-85233-896-1.ISSN 1431-875X.
  8. ^L. Castañeda; V. Arunachalam & S. Dharmaraja (2012).Introduction to Probability and Stochastic Processes with Applications. Wiley. p. 67.ISBN 9781118344941.
  9. ^Billingsley, Patrick (1995).Probability and Measure (3rd ed.). Wiley. p. 187.ISBN 9781466575592.
  10. ^abcdBertsekas, Dimitri P. (2002).Introduction to Probability. Tsitsiklis, John N., Τσιτσικλής, Γιάννης Ν. Belmont, Mass.: Athena Scientific.ISBN 188652940X.OCLC 51441829.
  11. ^Steigerwald, Douglas G."Economics 245A – Introduction to Measure Theory"(PDF). University of California, Santa Barbara. RetrievedApril 26, 2013.
  12. ^Fristedt & Gray (1996, page 11)

Literature

[edit]

External links

[edit]
Continuous data
Center
Dispersion
Shape
Count data
Summary tables
Dependence
Graphics
Study design
Survey methodology
Controlled experiments
Adaptive designs
Observational studies
Statistical theory
Frequentist inference
Point estimation
Interval estimation
Testing hypotheses
Parametric tests
Specific tests
Goodness of fit
Rank statistics
Bayesian inference
Correlation
Regression analysis
Linear regression
Non-standard predictors
Generalized linear model
Partition of variance
Categorical
Multivariate
Time-series
General
Specific tests
Time domain
Frequency domain
Survival
Survival function
Hazard function
Test
Biostatistics
Engineering statistics
Social statistics
Spatial statistics
Authority control databases: NationalEdit this at Wikidata
Retrieved from "https://en.wikipedia.org/w/index.php?title=Random_variable&oldid=1282623625"
Category:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp