Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Markov property

From Wikipedia, the free encyclopedia
Memoryless property of a stochastic process
This article is about the property of a stochastic process. For the class of properties of a finitely presented group, seeAdian–Rabin theorem.
A single realisation of three-dimensionalBrownian motion for times 0 ≤ t ≤ 2. Brownian motion has the Markov property, as the displacement of the particle does not depend on its past displacements.

Inprobability theory andstatistics, the termMarkov property refers to thememoryless property of astochastic process, which means that its future evolution is independent of its history. It is named after theRussianmathematicianAndrey Markov. The termstrong Markov property is similar to the Markov property, except that the meaning of "present" is defined in terms of arandom variable known as astopping time.

The termMarkov assumption is used to describe a model where the Markov property is assumed to hold, such as ahidden Markov model.

AMarkov random field extends this property to two or more dimensions or to random variables defined for an interconnected network of items.[1] An example of a model for such a field is theIsing model.

A discrete-time stochastic process satisfying the Markov property is known as aMarkov chain.

Introduction

[edit]

A stochastic process has the Markov property if theconditional probability distribution of future states of the process (conditional on both past and present values) depends only upon the present state; that is, given the present, the future does not depend on the past. A process with this property is said to beMarkov orMarkovian and known as aMarkov process. Two famous classes of Markov process are theMarkov chain andBrownian motion.

Note that there is a subtle, often overlooked and very important point that is often missed in the plain English statement of the definition: the statespace of the process is constant through time. The conditional description involves a fixed "bandwidth". For example, without this restriction we could augment any process to one which includes the complete history from a given initial condition and it would be made to be Markovian. But the state space would be of increasing dimensionality over time and does not meet the definition.

History

[edit]
Main article:Markov chain § History

Definition

[edit]

Let(Ω,F,P){\displaystyle (\Omega ,{\mathcal {F}},P)} be aprobability space with afiltration(Fs, sI){\displaystyle ({\mathcal {F}}_{s},\ s\in I)}, for some (totally ordered) index setI{\displaystyle I}; and let(S,Σ){\displaystyle (S,\Sigma )} be ameasurable space. An(S,Σ){\displaystyle (S,\Sigma )}-valued stochastic processX={Xt:ΩS}tI{\displaystyle X=\{X_{t}:\Omega \to S\}_{t\in I}}adapted to the filtration is said to possess theMarkov property if, for eachAΣ{\displaystyle A\in \Sigma } and eachs,tI{\displaystyle s,t\in I} withs<t{\displaystyle s<t},

P(XtAFs)=P(XtAXs).{\displaystyle P(X_{t}\in A\mid {\mathcal {F}}_{s})=P(X_{t}\in A\mid X_{s}).}[2]

In the case whereS{\displaystyle S} is a discrete set with thediscrete sigma algebra andI=N{\displaystyle I=\mathbb {N} }, this can be reformulated as follows:

P(Xn+1=xn+1Xn=xn,,X1=x1)=P(Xn+1=xn+1Xn=xn) for all nN.{\displaystyle P(X_{n+1}=x_{n+1}\mid X_{n}=x_{n},\dots ,X_{1}=x_{1})=P(X_{n+1}=x_{n+1}\mid X_{n}=x_{n}){\text{ for all }}n\in \mathbb {N} .}

In other words, the distribution ofX{\displaystyle X} at timen+1{\displaystyle n+1} depend solely on the state ofX{\displaystyle X} at timen{\displaystyle n} and is independent of the state of the process at any time previous ton{\displaystyle n}, which corresponds precisely to the intuition described in the introduction.

IfI=[0,){\displaystyle I=[0,\infty )}, thenX{\displaystyle X} is calledtime-homogeneous if for allt,s0{\displaystyle t,s\geq 0} the weak Markov property holds:[3]

P(Xt+sAFs)=P(XtAX0=x)|x=Xs=:PXs(XtA){\displaystyle P(X_{t+s}\in A\mid {\mathcal {F}}_{s})=P(X_{t}\in A\mid X_{0}=x)|_{x=X_{s}}=:P^{X_{s}}(X_{t}\in A)}.

The newly introduced probability measurePx(Xt){\displaystyle P^{x}(X_{t}\in \cdot )},xS{\displaystyle x\in S}, has the following intuition: It gives the probability that the processX{\displaystyle X} lies in some set at timet{\displaystyle t}, when it was started inx{\displaystyle x} at time zero. The functionPt(x,A):=Px(XtA){\displaystyle P_{t}(x,A):=P^{x}(X_{t}\in A)},(t,x,A)R+×S×Σ{\displaystyle (t,x,A)\in \mathbb {R} _{+}\times S\times \Sigma }, is also called thetransition function ofX{\displaystyle X} and the collection(Pt)t0{\displaystyle (P_{t})_{t\geq 0}} itstransition semigroup.

Alternative formulations

[edit]

There exists multiple alternative formulations of the elementary Markov property described above. The following are all equivalent:[4][5]

P(ABXt)=P(AXt)P(BXt){\displaystyle P(A\cap B\mid X_{t})=P(A\mid X_{t})P(B\mid X_{t})}.

P(BFt)=P(BXt){\displaystyle P(B\mid {\mathcal {F}}_{t})=P(B\mid X_{t})}.

P(AFt)=P(AXt){\displaystyle P(A\mid {\mathcal {F}}_{t}^{'})=P(A\mid X_{t})}.

E[YFt]=E[YXt]{\displaystyle \operatorname {E} [Y\mid {\mathcal {F}}_{t}]=\operatorname {E} [Y\mid X_{t}]}.

E[f(Xt)Fs]=E[f(Xt)Xs]{\displaystyle \operatorname {E} [f(X_{t})\mid {\mathcal {F}}_{s}]=\operatorname {E} [f(X_{t})\mid X_{s}]}.

E[f(Xt)Fs]=E[f(Xt)Xs]{\displaystyle \operatorname {E} [f(X_{t})\mid {\mathcal {F}}_{s}]=\operatorname {E} [f(X_{t})\mid X_{s}]}.

E[f(Xt)Xs,Xsn,...,Xs1]=E[f(Xt)Xs]{\displaystyle \operatorname {E} [f(X_{t})\mid X_{s},X_{s_{n}},...,X_{s_{1}}]=\operatorname {E} [f(X_{t})\mid X_{s}]}.

If there exists a so-calledshift-semigroup(θt)t0{\displaystyle (\theta _{t})_{t\geq 0}}, i.e., functionsθt:ΩΩ{\displaystyle \theta _{t}:\Omega \to \Omega } such that

  1. θ0=idΩ{\displaystyle \theta _{0}=\mathrm {id} _{\Omega }},
  2. θtθs=θt+ss,t0{\displaystyle \theta _{t}\circ \theta _{s}=\theta _{t+s}\quad \forall s,t\geq 0} (semigroup property),
  3. Xtθs=Xt+ss,t0{\displaystyle X_{t}\circ \theta _{s}=X_{t+s}\quad \forall s,t\geq 0},

then the Markov property is equivalent to:[4]

P(θt1(Λ)Ft)=P(θt1(Λ)Xt){\displaystyle P(\theta _{t}^{-1}(\Lambda )\mid {\mathcal {F}}_{t})=P(\theta _{t}^{-1}(\Lambda )\mid X_{t})}.

E[YθtFt]=E[YθtXt]{\displaystyle \operatorname {E} [Y\circ \theta _{t}\mid {\mathcal {F}}_{t}]=\operatorname {E} [Y\circ \theta _{t}\mid X_{t}]}.

Depending on the situation, some formulations might be easier to verify or to use than others.

Strong Markov property

[edit]

Suppose thatX=(Xt:t0){\displaystyle X=(X_{t}:t\geq 0)} is astochastic process on aprobability space(Ω,F,P){\displaystyle (\Omega ,{\mathcal {F}},P)} withnatural filtration{Ft}t0{\displaystyle \{{\mathcal {F}}_{t}\}_{t\geq 0}}. Then for anystopping timeτ{\displaystyle \tau } onΩ{\displaystyle \Omega }, we can define

Fτ={AF:t0,{τt}AFt}{\displaystyle {\mathcal {F}}_{\tau }=\{A\in {\mathcal {F}}:\forall t\geq 0,\{\tau \leq t\}\cap A\in {\mathcal {F}}_{t}\}}.

ThenX{\displaystyle X} is said to have the strong Markov property if, for eachstopping timeτ{\displaystyle \tau }, conditional on the event{τ<}{\displaystyle \{\tau <\infty \}}, we have that for eacht0{\displaystyle t\geq 0},Xτ+t{\displaystyle X_{\tau +t}} is independent ofFτ{\displaystyle {\mathcal {F}}_{\tau }} givenXτ{\displaystyle X_{\tau }}. This is equivalent to

P(Xτ+tA,τ<Fτ)=1{τ<}P(XtAX0=Xτ){\displaystyle P(X_{\tau +t}\in A,\tau <\infty \mid {\mathcal {F}}_{\tau })=1_{\{\tau <\infty \}}P(X_{t}\in A\mid X_{0}=X_{\tau })} for allAF{\displaystyle A\in {\mathcal {F}}},

where1{τ<}{\displaystyle 1_{\{\tau <\infty \}}} denotes to indicator function of the set{τ<}{\displaystyle \{\tau <\infty \}}.

The strong Markov property implies the ordinary Markov property since by taking the stopping timeτ=t{\displaystyle \tau =t}, the ordinary Markov property can be deduced.[6] The converse is in general not true.

The strong Markov property only leads to non-trivial results in continuous time (i.e., results which do not hold with merely the Markov property), as in the discrete case the strong and the elementary Markov property are equivalent.[7]

Feller property

[edit]
Main article:Feller process

Although the strong Markov property is in general stronger than the elementary Markov property, it is fulfilled by Markov processes with sufficiently "nice" regularity properties.

A continuous time Markov process is said to have theFeller property, if its transition semigroup(Pt)t0{\displaystyle (P_{t})_{t\geq 0}} (see above) fulfills[4]

  1. Ptf:=f(x)Pt(,dx)C0(S){\displaystyle P_{t}f:=\int f(x)P_{t}(\cdot ,dx)\in C_{0}(S)} for allfC0(S){\displaystyle f\in C_{0}(S)},
  2. limt0||Ptff||=0{\displaystyle \lim _{t\to 0}||P_{t}f-f||_{\infty }=0} for allfC0(S){\displaystyle f\in C_{0}(S)},

whereC0(S){\displaystyle C_{0}(S)} denotes the set ofcontinuous functionsvanishing at infinity and||||{\displaystyle ||\cdot ||_{\infty }} thesup norm. Then one can show that (if the filtration isaugmented) such a process has aversion with right-continuous (evencàdlàg) paths, which in turn fulfills the strong Markov property.

Examples

[edit]

Intuitive example

[edit]

Assume that an urn contains two red balls and one green ball. One ball was drawn yesterday, one ball was drawn today, and the final ball will be drawn tomorrow. All of the draws are "without replacement".

Suppose you know that today's ball was red, but you have no information about yesterday's ball. The chance that tomorrow's ball will be red is 1/2. That's because the only two remaining outcomes for this random experiment are:

DayOutcome 1Outcome 2
YesterdayRedGreen
TodayRedRed
TomorrowGreenRed

On the other hand, if you know that both today and yesterday's balls were red, then you are guaranteed to get a green ball tomorrow.

This discrepancy shows that the probability distribution for tomorrow's color depends not only on the present value, but is also affected by information about the past. This stochastic process of observed colors doesn't have the Markov property. Using the same experiment above, if sampling "without replacement" is changed to sampling "with replacement," the process of observed colors will have the Markov property.[8]

Stochastic processes

[edit]

Many prominent stochastic processes are Markov processes: TheBrownian motion, theBrownian bridge, thestochastic exponential, theOrnstein-Uhlenbeck process and thePoisson process have the Markov property.

More generally, any semimartingaleX{\displaystyle X} with values inRn{\displaystyle \mathbb {R} ^{n}} that is given by thestochastic differential equation

Xt=X0+i=1d[0tgi(Xs)ds+0tfi(Xs)dBsi]{\displaystyle X_{t}=X_{0}+\sum _{i=1}^{d}{\Big [}\int _{0}^{t}g_{i}(X_{s})ds+\int _{0}^{t}f_{i}(X_{s})dB_{s}^{i}{\Big ]}},

whereB=(B1,...,Bd){\displaystyle B=(B^{1},...,B^{d})} is ad{\displaystyle d}-dimensional Brownian motion andf1,...,fd,g1,...,gd:RnRn{\displaystyle f_{1},...,f_{d},g_{1},...,g_{d}:\mathbb {R} ^{n}\to \mathbb {R} ^{n}} are autonomous (i.e., they do not depend on time) Lipschitz functions, is time-homogeneous and has the strong Markov property. Iff1,...,fd,g1,...,gd{\displaystyle f_{1},...,f_{d},g_{1},...,g_{d}} are not autonomous, thenX{\displaystyle X} still has the elementary Markov property.[3]

Applications

[edit]

Forecasting

[edit]

In the fields ofpredictive modelling andprobabilistic forecasting, the Markov property is considered desirable since it may enable the reasoning and resolution of the problem that otherwise would not be possible to be resolved because of itsintractability. Such a model is known as aMarkov model.

Markov Chain Monte Carlo

[edit]

An application of the Markov property in a generalized form is inMarkov chain Monte Carlo computations in the context ofBayesian statistics.

See also

[edit]

References

[edit]
  1. ^Dodge, Yadolah. (2006)The Oxford Dictionary of Statistical Terms,Oxford University Press.ISBN 0-19-850994-4
  2. ^Durrett, Rick.Probability: Theory and Examples. Fourth Edition.Cambridge University Press, 2010.
  3. ^abProtter, Philip (1992).Stochastic Integration and Differential Equations (2nd ed.). Springer-Verlag Berlin Heidelberg. pp. 235–242.ISBN 978-3-662-02619-9.
  4. ^abcChung, Kai Lai; Walsh, John B. (2005).Markov Processes, Brownian Motion, and Time Symmetry (2nd ed.). Springer Science+Business Media. pp. 1–5,49–56.ISBN 978-0387-22026-0.
  5. ^Øksendal, Bernt K. (2003).Stochastic Differential Equations: An Introduction with Applications. Springer, Berlin.ISBN 3-540-04758-1.
  6. ^Ethier, Stewart N. andKurtz, Thomas G.Markov Processes: Characterization and Convergence. Wiley Series in Probability and Mathematical Statistics, 1986, p. 158.
  7. ^Klenke, Achim (2020).Probability Theory (3rd ed.). Springer Cham. p. 397.ISBN 978-3-030-56402-5.
  8. ^"Example of a stochastic process which does not have the Markov property".Stack Exchange. Retrieved2020-07-07.
Discrete time
Continuous time
Both
Fields and other
Time series models
Financial models
Actuarial models
Queueing models
Properties
Limit theorems
Inequalities
Tools
Disciplines
Retrieved from "https://en.wikipedia.org/w/index.php?title=Markov_property&oldid=1334895919"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp