Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Expected value

From Wikipedia, the free encyclopedia
Average value of a random variable
This article is about the term used in probability theory and statistics. For other uses, seeExpected value (disambiguation).
"E(X)" redirects here. For theex{\displaystyle e^{x}} function, seeExponential function.
Part of a series onstatistics
Probability theory

Inprobability theory, theexpected value (also calledexpectation,expectancy,expectation operator,mathematical expectation,mean,expectation value, orfirstmoment) is a generalization of theweighted average.

The expected value of a random variable with a finite number of outcomes is a weighted average of all possible outcomes. In the case of a continuum of possible outcomes, the expectation is defined byintegration. In the axiomatic foundation for probability provided bymeasure theory, the expectation is given byLebesgue integration.

The expected value of a random variableX is often denoted byE(X){\displaystyle {\text{E}}(X)},E[X]{\displaystyle {\text{E}}[X]}, orEX{\displaystyle {\text{E}}X}, withE also often stylized asE{\displaystyle \mathbb {E} } orE.[1][2][3]

History

[edit]

The idea of the expected value originated in the middle of the 17th century from the study of the so-calledproblem of points, which seeks to divide the stakesin a fair way between two players, who have to end their game before it is properly finished.[4] This problem had been debated for centuries. Many conflicting proposals and solutions had been suggested over the years when it was posed toBlaise Pascal by French writer and amateur mathematicianChevalier de Méré in 1654. Méré claimed that this problem could not be solved and that it showed just how flawed mathematics was when it came to its application to the real world. Pascal, being a mathematician, decided to work on a solution to the problem.

He began to discuss the problem in the famous series of letters toPierre de Fermat. Soon enough, they both independently came up with a solution. They solved the problem in different computational ways, but their results were identical because their computations were based on the same fundamental principle. The principle is that the value of a future gain should be directly proportional to the chance of getting it. This principle seemed to have come naturally to both of them. They were very pleased by the fact that they had found essentially the same solution, and this in turn made them absolutely convinced that they had solved the problem conclusively; however, they did not publish their findings. They only informed a small circle of mutual scientific friends in Paris about it.[5]

In Dutch mathematicianChristiaan Huygens' book, he considered the problem of points, and presented a solution based on the same principle as the solutions of Pascal and Fermat. Huygens published his treatise in 1657, (seeHuygens (1657)) "De ratiociniis in ludo aleæ" on probability theory just after visiting Paris. The book extended the concept of expectation by adding rules for how to calculate expectations in more complicated situations than the original problem (e.g., for three or more players), and can be seen as the first successful attempt at laying down the foundations of thetheory of probability.

In the foreword to his treatise, Huygens wrote:

It should be said, also, that for some time some of the best mathematicians of France have occupied themselves with this kind of calculus so that no one should attribute to me the honour of the first invention. This does not belong to me. But these savants, although they put each other to the test by proposing to each other many questions difficult to solve, have hidden their methods. I have had therefore to examine and go deeply for myself into this matter by beginning with the elements, and it is impossible for me for this reason to affirm that I have even started from the same principle. But finally I have found that my answers in many cases do not differ from theirs.

— Edwards (2002)

In the mid-nineteenth century,Pafnuty Chebyshev became the first person to think systematically in terms of the expectations ofrandom variables.[6]

Etymology

[edit]

Neither Pascal nor Huygens used the term "expectation" in its modern sense. In particular, Huygens writes:[7]

That any one Chance or Expectation to win any thing is worth just such a Sum, as wou'd procure in the same Chance and Expectation at a fair Lay. ... If I expect a or b, and have an equal chance of gaining them, my Expectation is worth (a+b)/2.

More than a hundred years later, in 1814,Pierre-Simon Laplace published his tract "Théorie analytique des probabilités", where the concept of expected value was defined explicitly:[8]

... this advantage in the theory of chance is the product of the sum hoped for by the probability of obtaining it; it is the partial sum which ought to result when we do not wish to run the risks of the event in supposing that the division is made proportional to the probabilities. This division is the only equitable one when all strange circumstances are eliminated; because an equal degree of probability gives an equal right for the sum hoped for. We will call this advantagemathematical hope.

Notations

[edit]

The use of the letterE to denote "expected value" goes back toW. A. Whitworth in 1901.[9] The symbol has since become popular for English writers. In German,E stands forErwartungswert, in Spanish foresperanza matemática, and in French forespérance mathématique.[10]

When "E" is used to denote "expected value", authors use a variety of stylizations: the expectation operator can be stylized asE (upright),E (italic), orE{\displaystyle \mathbb {E} } (inblackboard bold), while a variety of bracket notations (such asE(X),E[X], andEX) are all used.

Another popular notation isμX.X,Xav, andX¯{\displaystyle {\overline {X}}} are commonly used in physics.[11]M(X) is used in Russian-language literature.

Definition

[edit]

As discussed above, there are several context-dependent ways of defining the expected value. The simplest and original definition deals with the case of finitely many possible outcomes, such as in the flip of a coin. With the theory of infinite series, this can be extended to the case of countably many possible outcomes. It is also very common to consider the distinct case of random variables dictated by (piecewise-)continuousprobability density functions, as these arise in many natural contexts. All of these specific definitions may be viewed as special cases of the general definition based upon the mathematical tools ofmeasure theory andLebesgue integration, which provide these different contexts with an axiomatic foundation and common language.

Any definition of expected value may be extended to define an expected value of a multidimensional random variable, i.e. arandom vectorX. It is defined component by component, asE[X]i = E[Xi]. Similarly, one may define the expected value of arandom matrixX with componentsXij byE[X]ij = E[Xij].

Random variables with finitely many outcomes

[edit]

Consider a random variableX with afinite listx1, ...,xk of possible outcomes, each of which (respectively) has probabilityp1, ...,pk of occurring. The expectation ofX is defined as[12]E[X]=x1p1+x2p2++xkpk.{\displaystyle \operatorname {E} [X]=x_{1}p_{1}+x_{2}p_{2}+\cdots +x_{k}p_{k}.}

Since the probabilities must satisfyp1 + ⋅⋅⋅ +pk = 1, it is natural to interpretE[X] as aweighted average of thexi values, with weights given by their probabilitiespi.

In the special case that all possible outcomes areequiprobable (that is,p1 = ⋅⋅⋅ =pk), the weighted average is given by the standardaverage. In the general case, the expected value takes into account the fact that some outcomes are more likely than others.

Examples

[edit]
An illustration of the convergence of sequence averages of rolls of a dice to the expected value of 3.5 as the number of rolls (trials) grows

Random variables with countably infinitely many outcomes

[edit]

Informally, the expectation of a random variable with acountably infinite set of possible outcomes is defined analogously as the weighted average of all possible outcomes, where the weights are given by the probabilities of realizing each given value. This is to say thatE[X]=i=1xipi,{\displaystyle \operatorname {E} [X]=\sum _{i=1}^{\infty }x_{i}\,p_{i},}wherex1,x2, ... are the possible outcomes of the random variableX andp1,p2, ... are their corresponding probabilities. In many non-mathematical textbooks, this is presented as the full definition of expected values in this context.[13]

However, there are some subtleties with infinite summation, so the above formula is not suitable as a mathematical definition. In particular, theRiemann series theorem ofmathematical analysis illustrates that the value of certain infinite sums involving positive and negative summands depends on the order in which the summands are given. Since the outcomes of a random variable have no naturally given order, this creates a difficulty in defining expected value precisely.

For this reason, many mathematical textbooks only consider the case that the infinite sum given aboveconverges absolutely, which implies that the infinite sum is a finite number independent of the ordering of summands.[14] In the alternative case that the infinite sum does not converge absolutely, one says the random variabledoes not have finite expectation.[14]

Examples

[edit]

Random variables with density

[edit]

Now consider a random variableX which has aprobability density function given by a functionf on thereal number line. This means that the probability ofX taking on a value in any givenopen interval is given by theintegral off over that interval. The expectation ofX is then given by the integral[15]E[X]=xf(x)dx.{\displaystyle \operatorname {E} [X]=\int _{-\infty }^{\infty }xf(x)\,dx.}A general and mathematically precise formulation of this definition usesmeasure theory andLebesgue integration, and the corresponding theory ofabsolutely continuous random variables is described in the next section. The density functions of many common distributions arepiecewise continuous, and as such the theory is often developed in this restricted setting.[16] For such functions, it is sufficient to only consider the standardRiemann integration. Sometimescontinuous random variables are defined as those corresponding to this special class of densities, although the term is used differently by various authors.

Analogously to the countably-infinite case above, there are subtleties with this expression due to the infinite region of integration. Such subtleties can be seen concretely if the distribution ofX is given by theCauchy distributionCauchy(0, π), so thatf(x) = (x2 + π2)−1. It is straightforward to compute in this case thatabxf(x)dx=abxx2+π2dx=12lnb2+π2a2+π2.{\displaystyle \int _{a}^{b}xf(x)\,dx=\int _{a}^{b}{\frac {x}{x^{2}+\pi ^{2}}}\,dx={\frac {1}{2}}\ln {\frac {b^{2}+\pi ^{2}}{a^{2}+\pi ^{2}}}.}The limit of this expression asa → −∞ andb → ∞ does not exist: if the limits are taken so thata = −b, then the limit is zero, while if the constraint2a = −b is taken, then the limit isln(2).

To avoid such ambiguities, in mathematical textbooks it is common to require that the given integralconverges absolutely, withE[X] left undefined otherwise.[17] However, measure-theoretic notions as given below can be used to give a systematic definition ofE[X] for more general random variablesX.

Arbitrary real-valued random variables

[edit]

All definitions of the expected value may be expressed in the language ofmeasure theory. In general, ifX is a real-valuedrandom variable defined on aprobability space(Ω, Σ, P), then the expected value ofX, denoted byE[X], is defined as theLebesgue integral[18]E[X]=ΩXdP.{\displaystyle \operatorname {E} [X]=\int _{\Omega }X\,d\operatorname {P} .}Despite the newly abstract situation, this definition is extremely similar in nature to the very simplest definition of expected values, given above, as certain weighted averages. This is because, in measure theory, the value of the Lebesgue integral ofX is defined via weighted averages ofapproximations ofX which take on finitely many values.[19] Moreover, if given a random variable with finitely or countably many possible values, the Lebesgue theory of expectation is identical to the summation formulas given above. However, the Lebesgue theory clarifies the scope of the theory of probability density functions. A random variableX is said to beabsolutely continuous if any of the following conditions are satisfied:

These conditions are all equivalent, although this is nontrivial to establish.[20] In this definition,f is called theprobability density function ofX (relative to Lebesgue measure). According to the change-of-variables formula for Lebesgue integration,[21] combined with thelaw of the unconscious statistician,[22] it follows thatE[X]ΩXdP=Rxf(x)dx{\displaystyle \operatorname {E} [X]\equiv \int _{\Omega }X\,d\operatorname {P} =\int _{\mathbb {R} }xf(x)\,dx}for any absolutely continuous random variableX. The above discussion of continuous random variables is thus a special case of the general Lebesgue theory, due to the fact that every piecewise-continuous function is measurable.

Expected value μ and median 𝑚
Expected valueμ and median𝑚

The expected value of any real-valued random variableX{\displaystyle X} can also be defined on the graph of itscumulative distribution functionF{\displaystyle F} by a nearby equality of areas. In fact,E[X]=μ{\displaystyle \operatorname {E} [X]=\mu } with a real numberμ{\displaystyle \mu } if and only if the two surfaces in thex{\displaystyle x}-y{\displaystyle y}-plane, described byxμ,0yF(x)orxμ,F(x)y1{\displaystyle x\leq \mu ,\;\,0\leq y\leq F(x)\quad {\text{or}}\quad x\geq \mu ,\;\,F(x)\leq y\leq 1}respectively, have the same finite area, i.e. ifμF(x)dx=μ(1F(x))dx{\displaystyle \int _{-\infty }^{\mu }F(x)\,dx=\int _{\mu }^{\infty }{\big (}1-F(x){\big )}\,dx}and bothimproper Riemann integrals converge. Finally, this is equivalent to the representationE[X]=0(1F(x))dx0F(x)dx,{\displaystyle \operatorname {E} [X]=\int _{0}^{\infty }{\bigl (}1-F(x){\bigr )}\,dx-\int _{-\infty }^{0}F(x)\,dx,}also with convergent integrals.[23]

Infinite expected values

[edit]

Expected values as defined above are automatically finite numbers. However, in many cases it is fundamental to be able to consider expected values of±∞. This is intuitive, for example, in the case of theSt. Petersburg paradox, in which one considers a random variable with possible outcomesxi = 2i, with associated probabilitiespi = 2i, fori ranging over all positive integers. According to the summation formula in the case of random variables with countably many outcomes, one hasE[X]=i=1xipi=212+414+818+16116+=1+1+1+1+.{\displaystyle \operatorname {E} [X]=\sum _{i=1}^{\infty }x_{i}\,p_{i}=2\cdot {\frac {1}{2}}+4\cdot {\frac {1}{4}}+8\cdot {\frac {1}{8}}+16\cdot {\frac {1}{16}}+\cdots =1+1+1+1+\cdots .}It is natural to say that the expected value equals+∞.

There is a rigorous mathematical theory underlying such ideas, which is often taken as part of the definition of the Lebesgue integral.[19] The first fundamental observation is that, whichever of the above definitions are followed, anynonnegative random variable whatsoever can be given an unambiguous expected value; whenever absolute convergence fails, then the expected value can be defined as+∞. The second fundamental observation is that any random variable can be written as the difference of two nonnegative random variables. Given a random variableX, one defines thepositive and negative parts byX + = max(X, 0) andX = −min(X, 0). These are nonnegative random variables, and it can be directly checked thatX =X +X. SinceE[X +] andE[X] are both then defined as either nonnegative numbers or+∞, it is then natural to define:E[X]={E[X+]E[X]if E[X+]< and E[X]<;+if E[X+]= and E[X]<;if E[X+]< and E[X]=;undefinedif E[X+]= and E[X]=.{\displaystyle \operatorname {E} [X]={\begin{cases}\operatorname {E} [X^{+}]-\operatorname {E} [X^{-}]&{\text{if }}\operatorname {E} [X^{+}]<\infty {\text{ and }}\operatorname {E} [X^{-}]<\infty ;\\+\infty &{\text{if }}\operatorname {E} [X^{+}]=\infty {\text{ and }}\operatorname {E} [X^{-}]<\infty ;\\-\infty &{\text{if }}\operatorname {E} [X^{+}]<\infty {\text{ and }}\operatorname {E} [X^{-}]=\infty ;\\{\text{undefined}}&{\text{if }}\operatorname {E} [X^{+}]=\infty {\text{ and }}\operatorname {E} [X^{-}]=\infty .\end{cases}}}

According to this definition,E[X] exists and is finite if and only ifE[X +] andE[X] are both finite. Due to the formula|X| =X + +X, this is the case if and only ifE|X| is finite, and this is equivalent to the absolute convergence conditions in the definitions above. As such, the present considerations do not define finite expected values in any cases not previously considered; they are only useful for infinite expectations.

  • In the case of the St. Petersburg paradox, one hasX = 0 and soE[X] = +∞ as desired.
  • Suppose the random variableX takes values1, −2,3, −4, ... with respective probabilities−2, 6(2π)−2, 6(3π)−2, 6(4π)−2, .... Then it follows thatX + takes value2k−1 with probability6((2k−1)π)−2 for each positive integerk, and takes value0 with remaining probability. Similarly,X takes value2k with probability6(2kπ)−2 for each positive integerk and takes value0 with remaining probability. Using the definition for non-negative random variables, one can show that bothE[X +] = ∞ andE[X] = ∞ (seeHarmonic series). Hence, in this case the expectation ofX is undefined.
  • Similarly, the Cauchy distribution, as discussed above, has undefined expectation.

Tail-sum formula

[edit]

In the case of a non-negative integer-valued random variable X, the expected value can also be expressed in terms of itstail probabilities (sometimes called thetail-sum formula):

E[X]=k=0Pr(X>k).{\displaystyle \operatorname {E} [X]=\sum _{k=0}^{\infty }\Pr(X>k).}

A more general version holds for any non-negative random variable (discrete or continuous):

E[X]=0Pr(X>t)dt,{\displaystyle \operatorname {E} [X]=\int _{0}^{\infty }\Pr(X>t)\,dt,}

where the integrand is the survival function of X.


Expected values of common distributions

[edit]

The following table gives the expected values of some commonly occurringprobability distributions. The third column gives the expected values both in the form immediately given by the definition, as well as in the simplified form obtained by computation therefrom. The details of these computations, which are not always straightforward, can be found in the indicated references.

DistributionNotationMean E(X)
Bernoulli[24]X b(1,p){\displaystyle X\sim ~b(1,p)}0(1p)+1p=p{\displaystyle 0\cdot (1-p)+1\cdot p=p}
Binomial[25]XB(n,p){\displaystyle X\sim B(n,p)}i=0ni(ni)pi(1p)ni=np{\displaystyle \sum _{i=0}^{n}i{n \choose i}p^{i}(1-p)^{n-i}=np}
Poisson[26]XPo(λ){\displaystyle X\sim \mathrm {Po} (\lambda )}i=0ieλλii!=λ{\displaystyle \sum _{i=0}^{\infty }{\frac {ie^{-\lambda }\lambda ^{i}}{i!}}=\lambda }
Geometric[27]XGeometric(p){\displaystyle X\sim \mathrm {Geometric} (p)}i=1ip(1p)i1=1p{\displaystyle \sum _{i=1}^{\infty }ip(1-p)^{i-1}={\frac {1}{p}}}
Uniform[28]XU(a,b){\displaystyle X\sim U(a,b)}abxbadx=a+b2{\displaystyle \int _{a}^{b}{\frac {x}{b-a}}\,dx={\frac {a+b}{2}}}
Exponential[29]Xexp(λ){\displaystyle X\sim \exp(\lambda )}0λxeλxdx=1λ{\displaystyle \int _{0}^{\infty }\lambda xe^{-\lambda x}\,dx={\frac {1}{\lambda }}}
Normal[30]XN(μ,σ2){\displaystyle X\sim N(\mu ,\sigma ^{2})}12πσ2xe12(xμσ)2dx=μ{\displaystyle {\frac {1}{\sqrt {2\pi \sigma ^{2}}}}\int _{-\infty }^{\infty }x\,e^{-{\frac {1}{2}}\left({\frac {x-\mu }{\sigma }}\right)^{2}}\,dx=\mu }
Standard Normal[31]XN(0,1){\displaystyle X\sim N(0,1)}12πxex2/2dx=0{\displaystyle {\frac {1}{\sqrt {2\pi }}}\int _{-\infty }^{\infty }xe^{-x^{2}/2}\,dx=0}
Pareto[32]XPar(α,k){\displaystyle X\sim \mathrm {Par} (\alpha ,k)}kαkαxαdx={αkα1if α>1if 0<α1{\displaystyle \int _{k}^{\infty }\alpha k^{\alpha }x^{-\alpha }\,dx={\begin{cases}{\frac {\alpha k}{\alpha -1}}&{\text{if }}\alpha >1\\\infty &{\text{if }}0<\alpha \leq 1\end{cases}}}
Cauchy[33]XCauchy(x0,γ){\displaystyle X\sim \mathrm {Cauchy} (x_{0},\gamma )}1πγx(xx0)2+γ2dx{\displaystyle {\frac {1}{\pi }}\int _{-\infty }^{\infty }{\frac {\gamma x}{(x-x_{0})^{2}+\gamma ^{2}}}\,dx} isundefined

Properties

[edit]

The basic properties below (and their names in bold) replicate or follow immediately from those ofLebesgue integral. Note that the letters "a.s." stand for "almost surely"—a central property of the Lebesgue integral. Basically, one says that an inequality likeX0{\displaystyle X\geq 0} is true almost surely, when the probability measure attributes zero-mass to the complementary event{X<0}.{\displaystyle \left\{X<0\right\}.}

Inequalities

[edit]

Concentration inequalities control the likelihood of a random variable taking on large values.Markov's inequality is among the best-known and simplest to prove: for anonnegative random variableX and any positive numbera, it states that[37]P(Xa)E[X]a.{\displaystyle \operatorname {P} (X\geq a)\leq {\frac {\operatorname {E} [X]}{a}}.}

IfX is any random variable with finite expectation, then Markov's inequality may be applied to the random variable|X−E[X]|2 to obtainChebyshev's inequalityP(|XE[X]|a)Var[X]a2,{\displaystyle \operatorname {P} (|X-{\text{E}}[X]|\geq a)\leq {\frac {\operatorname {Var} [X]}{a^{2}}},}whereVar is thevariance.[37] These inequalities are significant for their nearly complete lack of conditional assumptions. For example, for any random variable with finite expectation, the Chebyshev inequality implies that there is at least a 75% probability of an outcome being within twostandard deviations of the expected value. However, in special cases the Markov and Chebyshev inequalities often give much weaker information than is otherwise available. For example, in the case of an unweighted dice, Chebyshev's inequality says that odds of rolling between 1 and 6 is at least 53%; in reality, the odds are of course 100%.[38] TheKolmogorov inequality extends the Chebyshev inequality to the context of sums of random variables.[39]

The following three inequalities are of fundamental importance in the field ofmathematical analysis and its applications to probability theory.

The Hölder and Minkowski inequalities can be extended to generalmeasure spaces, and are often given in that context. By contrast, the Jensen inequality is special to the case of probability spaces.

Expectations under convergence of random variables

[edit]

In general, it is not the case thatE[Xn]E[X]{\displaystyle \operatorname {E} [X_{n}]\to \operatorname {E} [X]} even ifXnX{\displaystyle X_{n}\to X} pointwise. Thus, one cannot interchange limits and expectation, without additional conditions on the random variables. To see this, letU{\displaystyle U} be a random variable distributed uniformly on[0,1].{\displaystyle [0,1].} Forn1,{\displaystyle n\geq 1,} define a sequence of random variablesXn=n1{U(0,1n)},{\displaystyle X_{n}=n\cdot \mathbf {1} \left\{U\in \left(0,{\tfrac {1}{n}}\right)\right\},}with1{A}{\displaystyle \mathbf {1} \{A\}} being the indicator function of the eventA.{\displaystyle A.} Then, it follows thatXn0{\displaystyle X_{n}\to 0} pointwise. But,E[Xn]=nPr(U[0,1n])=n1n=1{\displaystyle \operatorname {E} [X_{n}]=n\cdot \Pr \left(U\in \left[0,{\tfrac {1}{n}}\right]\right)=n\cdot {\tfrac {1}{n}}=1} for eachn.{\displaystyle n.} Hence,limnE[Xn]=10=E[limnXn].{\displaystyle \lim _{n\to \infty }\operatorname {E} [X_{n}]=1\neq 0=\operatorname {E} \left[\lim _{n\to \infty }X_{n}\right].}

Analogously, for general sequence of random variables{Yn:n0},{\displaystyle \{Y_{n}:n\geq 0\},} the expected value operator is notσ{\displaystyle \sigma }-additive, i.e.E[n=0Yn]n=0E[Yn].{\displaystyle \operatorname {E} \left[\sum _{n=0}^{\infty }Y_{n}\right]\neq \sum _{n=0}^{\infty }\operatorname {E} [Y_{n}].}

An example is easily obtained by settingY0=X1{\displaystyle Y_{0}=X_{1}} andYn=Xn+1Xn{\displaystyle Y_{n}=X_{n+1}-X_{n}} forn1,{\displaystyle n\geq 1,} whereXn{\displaystyle X_{n}} is as in the previous example.

A number of convergence results specify exact conditions which allow one to interchange limits and expectations, as specified below.

Relationship with characteristic function

[edit]

The probability density functionfX{\displaystyle f_{X}} of a scalar random variableX{\displaystyle X} is related to itscharacteristic functionφX{\displaystyle \varphi _{X}} by the inversion formula:fX(x)=12πReitxφX(t)dt.{\displaystyle f_{X}(x)={\frac {1}{2\pi }}\int _{\mathbb {R} }e^{-itx}\varphi _{X}(t)\,dt.}

For the expected value ofg(X){\displaystyle g(X)} (whereg:RR{\displaystyle g:{\mathbb {R} }\to {\mathbb {R} }} is aBorel function), we can use this inversion formula to obtainE[g(X)]=12πRg(x)[ReitxφX(t)dt]dx.{\displaystyle \operatorname {E} [g(X)]={\frac {1}{2\pi }}\int _{\mathbb {R} }g(x)\left[\int _{\mathbb {R} }e^{-itx}\varphi _{X}(t)\,dt\right]dx.}

IfE[g(X)]{\displaystyle \operatorname {E} [g(X)]} is finite, changing the order of integration, we get, in accordance withFubini–Tonelli theorem,E[g(X)]=12πRG(t)φX(t)dt,{\displaystyle \operatorname {E} [g(X)]={\frac {1}{2\pi }}\int _{\mathbb {R} }G(t)\varphi _{X}(t)\,dt,}whereG(t)=Rg(x)eitxdx{\displaystyle G(t)=\int _{\mathbb {R} }g(x)e^{-itx}\,dx}is theFourier transform ofg(x).{\displaystyle g(x).} The expression forE[g(X)]{\displaystyle \operatorname {E} [g(X)]} also follows directly from thePlancherel theorem.

Uses and applications

[edit]

The expectation of a random variable plays an important role in a variety of contexts.

Instatistics, where one seeksestimates for unknownparameters based on available data gained fromsamples, thesample mean serves as an estimate for the expectation, and is itself a random variable. In such settings, the sample mean is considered to meet the desirable criterion for a "good" estimator in beingunbiased; that is, the expected value of the estimate is equal to thetrue value of the underlying parameter.

See also:Estimation theory

For a different example, indecision theory, an agent making an optimal choice in the context of incomplete information is often assumed to maximize the expected value of theirutility function.

It is possible to construct an expected value equal to the probability of an event by taking the expectation of anindicator function that is one if the event has occurred and zero otherwise. This relationship can be used to translate properties of expected values into properties of probabilities, e.g. using thelaw of large numbers to justify estimating probabilities byfrequencies.

The expected values of the powers ofX are called themoments ofX; themoments about the mean ofX are expected values of powers ofX − E[X]. The moments of some random variables can be used to specify their distributions, via theirmoment generating functions.

To empirically estimate the expected value of a random variable, one repeatedly measures observations of the variable and computes thearithmetic mean of the results. If the expected value exists, this procedure estimates the true expected value in an unbiased manner and has the property of minimizing the sum of the squares of theresiduals (the sum of the squared differences between the observations and the estimate). The law of large numbers demonstrates (under fairly mild conditions) that, as thesize of the sample gets larger, thevariance of this estimate gets smaller.

This property is often exploited in a wide variety of applications, including general problems ofstatistical estimation andmachine learning, to estimate (probabilistic) quantities of interest viaMonte Carlo methods, since most quantities of interest can be written in terms of expectation, e.g.P(XA)=E[1A],{\displaystyle \operatorname {P} ({X\in {\mathcal {A}}})=\operatorname {E} [{\mathbf {1} }_{\mathcal {A}}],} where1A{\displaystyle {\mathbf {1} }_{\mathcal {A}}} is the indicator function of the setA.{\displaystyle {\mathcal {A}}.}

The mass of probability distribution is balanced at the expected value, here a Beta(α,β) distribution with expected value α/(α+β).

Inclassical mechanics, thecenter of mass is an analogous concept to expectation. For example, supposeX is a discrete random variable with valuesxi and corresponding probabilitiespi. Now consider a weightless rod on which are placed weights, at locationsxi along the rod and having massespi (whose sum is one). The point at which the rod balances is E[X].

Expected values can also be used to compute the variance, by means of the computational formula for the varianceVar(X)=E[X2](E[X])2.{\displaystyle \operatorname {Var} (X)=\operatorname {E} [X^{2}]-(\operatorname {E} [X])^{2}.}

A very important application of the expectation value is in the field ofquantum mechanics. Theexpectation value of a quantum mechanical operatorA^{\displaystyle {\hat {A}}} operating on aquantum state vector|ψ{\displaystyle |\psi \rangle } is written asA^=ψ|A^|ψ.{\displaystyle \langle {\hat {A}}\rangle =\langle \psi |{\hat {A}}|\psi \rangle .} Theuncertainty inA^{\displaystyle {\hat {A}}} can be calculated by the formula(ΔA)2=A^2A^2{\displaystyle (\Delta A)^{2}=\langle {\hat {A}}^{2}\rangle -\langle {\hat {A}}\rangle ^{2}}.

See also

[edit]

References

[edit]
  1. ^"Expectation | Mean | Average".www.probabilitycourse.com. Retrieved2020-09-11.
  2. ^Hansen, Bruce."PROBABILITY AND STATISTICS FOR ECONOMISTS"(PDF). Archived fromthe original(PDF) on 2022-01-19. Retrieved2021-07-20.
  3. ^Wasserman, Larry (December 2010).All of Statistics: A Concise Course in Statistical Inference. Springer texts in statistics. p. 47.ISBN 9781441923226.
  4. ^Hald, Anders (1990).History of Probability and Statistics and Their Applications before 1750. Wiley Series in Probability and Statistics.doi:10.1002/0471725161.ISBN 9780471725169.
  5. ^Ore, Øystein (1960). "Ore, Pascal and the Invention of Probability Theory".The American Mathematical Monthly.67 (5):409–419.doi:10.2307/2309286.JSTOR 2309286.
  6. ^Mackey, George (July 1980). "HARMONIC ANALYSIS AS THE EXPLOITATION OF SYMMETRY - A HISTORICAL SURVEY".Bulletin of the American Mathematical Society. New Series.3 (1): 549.
  7. ^Huygens, Christian."The Value of Chances in Games of Fortune. English Translation"(PDF).
  8. ^Laplace, Pierre-Simon (1952) [1951].A philosophical essay on probabilities. Dover Publications.OCLC 475539.
  9. ^Whitworth, W.A. (1901)Choice and Chance with One Thousand Exercises. Fifth edition. Deighton Bell, Cambridge. [Reprinted by Hafner Publishing Co., New York, 1959.]
  10. ^"Earliest uses of symbols in probability and statistics".
  11. ^Feller 1968, p. 221.
  12. ^Billingsley 1995, p. 76.
  13. ^Ross 2019, Section 2.4.1.
  14. ^abFeller 1968, Section IX.2.
  15. ^Papoulis & Pillai 2002, Section 5-3;Ross 2019, Section 2.4.2.
  16. ^Feller 1971, Section I.2.
  17. ^Feller 1971, p. 5.
  18. ^Billingsley 1995, p. 273.
  19. ^abBillingsley 1995, Section 15.
  20. ^Billingsley 1995, Theorems 31.7 and 31.8 and p. 422.
  21. ^Billingsley 1995, Theorem 16.13.
  22. ^Billingsley 1995, Theorem 16.11.
  23. ^Uhl, Roland (2023).Charakterisierung des Erwartungswertes am Graphen der Verteilungsfunktion [Characterization of the expected value on the graph of the cumulative distribution function](PDF). Technische Hochschule Brandenburg. pp. 2–4.doi:10.25933/opus4-2986.Archived from the original on 2023-12-24.
  24. ^Casella & Berger 2001, p. 89;Ross 2019, Example 2.16.
  25. ^Casella & Berger 2001, Example 2.2.3;Ross 2019, Example 2.17.
  26. ^Billingsley 1995, Example 21.4;Casella & Berger 2001, p. 92;Ross 2019, Example 2.19.
  27. ^Casella & Berger 2001, p. 97;Ross 2019, Example 2.18.
  28. ^Casella & Berger 2001, p. 99;Ross 2019, Example 2.20.
  29. ^Billingsley 1995, Example 21.3;Casella & Berger 2001, Example 2.2.2;Ross 2019, Example 2.21.
  30. ^Casella & Berger 2001, p. 103;Ross 2019, Example 2.22.
  31. ^Billingsley 1995, Example 21.1;Casella & Berger 2001, p. 103.
  32. ^Johnson, Kotz & Balakrishnan 1994, Chapter 20.
  33. ^Feller 1971, Section II.4.
  34. ^abcWeisstein, Eric W."Expectation Value".mathworld.wolfram.com. Retrieved2020-09-11.
  35. ^Feller 1971, Section V.6.
  36. ^Papoulis & Pillai 2002, Section 6-4.
  37. ^abFeller 1968, Section IX.6;Feller 1971, Section V.7;Papoulis & Pillai 2002, Section 5-4;Ross 2019, Section 2.8.
  38. ^Feller 1968, Section IX.6.
  39. ^Feller 1968, Section IX.7.
  40. ^abcdFeller 1971, Section V.8.
  41. ^Billingsley 1995, pp. 81, 277.
  42. ^Billingsley 1995, Section 19.

Bibliography

[edit]
International
Other
Retrieved from "https://en.wikipedia.org/w/index.php?title=Expected_value&oldid=1322686115"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp