Inmathematics andstatistics, astationary process (also called astrict/strictly stationary process orstrong/strongly stationary process) is astochastic process whose statistical properties, such asmean andvariance, do not change over time. More formally, thejoint probability distribution of the process remains the same when shifted in time. This implies that the process is statistically consistent across different time periods. Because many statistical procedures intime series analysis assume stationarity, non-stationary data are frequently transformed to achieve stationarity before analysis.
A common cause of non-stationarity is a trend in the mean, which can be due to either aunit root or a deterministic trend. In the case of a unit root, stochastic shocks have permanent effects, and the process is notmean-reverting. With a deterministic trend, the process is calledtrend-stationary, and shocks have only transitory effects, with the variable tending towards a deterministically evolving mean. A trend-stationary process is not strictly stationary but can be made stationary by removing the trend. Similarly, processes with unit roots can be made stationary throughdifferencing.
Another type of non-stationary process, distinct from those with trends, is acyclostationary process, which exhibits cyclical variations over time.
Strict stationarity, as defined above, can be too restrictive for many applications. Therefore, other forms of stationarity, such aswide-sense stationarity orN-th-order stationarity, are often used. The definitions for different kinds of stationarity are not consistent among different authors (seeOther terminology).
Formally, let be astochastic process and let represent thecumulative distribution function of theunconditional (i.e., with no reference to any particular starting value)joint distribution of at times. Then, is said to bestrictly stationary,strongly stationary orstrict-sense stationary if[1]: p. 155
| Eq.1 |
Since does not affect, is independent of time.

White noise is the simplest example of a stationary process.
An example of adiscrete-time stationary process where the sample space is also discrete (so that the random variable may take one ofN possible values) is aBernoulli scheme. Other examples of a discrete-time stationary process with continuous sample space include someautoregressive andmoving average processes which are both subsets of theautoregressive moving average model. Models with a non-trivial autoregressive component may be either stationary or non-stationary, depending on the parameter values, and important non-stationary special cases are whereunit roots exist in the model.
Let be any scalarrandom variable, and define a time-series by
Then is a stationary time series, for which realisations consist of a series of constant values, with a different constant value for each realisation. Alaw of large numbers does not apply on this case, as the limiting value of an average from a single realisation takes the random value determined by, rather than taking theexpected value of.
The time average of does not converge since the process is notergodic.
As a further example of a stationary process for which any single realisation has an apparently noise-free structure, let have auniform distribution on and define the time series by
Then is strictly stationary since ( modulo) follows the same uniform distribution as for any.
Keep in mind that aweakly white noise is not necessarily strictly stationary. Let be a random variable uniformly distributed in the interval and define the time series
Then
So is a white noise in the weak sense (the mean and cross-covariances are zero, and the variances are all the same), however it is not strictly stationary.
InEq.1, the distribution of samples of the stochastic process must be equal to the distribution of the samples shifted in timefor all.N-th-order stationarity is a weaker form of stationarity where this is only requested for all up to a certain order. A random process is said to beN-th-order stationary if:[1]: p. 152
| Eq.2 |
A weaker form of stationarity commonly employed insignal processing is known asweak-sense stationarity,wide-sense stationarity (WSS), orcovariance stationarity. WSS random processes only require that 1stmoment (i.e. the mean) andautocovariance do not vary with respect to time and that the 2nd moment is finite for all times. Any strictly stationary process which has a finitemean andcovariance is also WSS.[2]: p. 299
So, acontinuous timerandom process which is WSS has the following restrictions on its mean function andautocovariance function:
| Eq.3 |
The first property implies that the mean function must be constant. The second property implies that the autocovariance function depends only on thedifference between and and only needs to be indexed by one variable rather than two variables.[1]: p. 159 Thus, instead of writing,
the notation is often abbreviated by the substitution:
This also implies that theautocorrelation depends only on, that is
The third property says that the second moments must be finite for any time.
The main advantage of wide-sense stationarity is that it places the time-series in the context ofHilbert spaces. LetH be the Hilbert space generated by {x(t)} (that is, the closure of the set of all linear combinations of these random variables in the Hilbert space of all square-integrable random variables on the given probability space). By the positive definiteness of the autocovariance function, it follows fromBochner's theorem that there exists a positive measure on the real line such thatH is isomorphic to the Hilbert subspace ofL2(μ) generated by {e−2πiξ⋅t}. This then gives the following Fourier-type decomposition for a continuous time stationary stochastic process: there exists a stochastic process withorthogonal increments such that, for all
where the integral on the right-hand side is interpreted in a suitable (Riemann) sense. The same result holds for a discrete-time stationary process, with the spectral measure now defined on the unit circle.
When processing WSS random signals withlinear,time-invariant (LTI)filters, it is helpful to think of the correlation function as alinear operator. Since it is acirculant operator (depends only on the difference between the two arguments), its eigenfunctions are theFourier complex exponentials. Additionally, since theeigenfunctions of LTI operators are alsocomplex exponentials, LTI processing of WSS random signals is highly tractable—all computations can be performed in thefrequency domain. Thus, the WSS assumption is widely employed in signal processingalgorithms.
In the case where is a complex stochastic process theautocovariance function is defined as and, in addition to the requirements inEq.3, it is required that the pseudo-autocovariance function depends only on the time lag. In formulas, is WSS, if
| Eq.4 |
The concept of stationarity may be extended to two stochastic processes.
Two stochastic processes and are calledjointly strict-sense stationary if their joint cumulative distribution remains unchanged under time shifts, i.e. if
| Eq.5 |
Two random processes and is said to bejointly (M + N)-th-order stationary if:[1]: p. 159
| Eq.6 |
Two stochastic processes and are calledjointly wide-sense stationary if they are both wide-sense stationary and their cross-covariance function depends only on the time difference. This may be summarized as follows:
| Eq.7 |
The terminology used for types of stationarity other than strict stationarity can be rather mixed. Some examples follow.
In time series analysis and stochastic processes, stationarizing a time series is a crucial preprocessing step aimed at transforming a non-stationary process into a stationary one. Several techniques exist for achieving this, depending on the type and order of non-stationarity present. For first-order non-stationarity, where the mean of the process varies over time, differencing is a common and effective method: it transforms the series by subtracting each value from its predecessor, thus stabilizing the mean. For non-stationarities up to the second order, time-frequency analysis (e.g.,Wavelet transform,Wigner distribution function, orShort-time Fourier transform) can be employed to isolate and suppress time-localized, nonstationary spectral components. Additionally, surrogate data methods can be used to construct strictly stationary versions of the original time series. One of the ways for identifying non-stationary times series is theACF plot. Sometimes, patterns will be more visible in the ACF plot than in the original time series; however, this is not always the case.[6]
The choice of method for time series stationarization depends on the nature of the non-stationarity and the goals of the analysis, especially when building models that require strict stationarity assumptions, such as ARMA or spectral-based techniques. More details on some time series stationarization methods are presented below.
One way to make some time series first-order stationary is to compute the differences between consecutive observations. This is known asdifferencing. Differencing can help stabilize the mean of a time series by removing changes in the level of a time series, and so eliminating trends. This can also remove seasonality, if differences are taken appropriately (e.g. differencing observations 1 year apart to remove a yearly trend). Transformations such as logarithms can help to stabilize the variance of a time series.
The surrogate method for stationarization[7] works by generating a new time series that preserves certain statistical properties of the original series while removing its nonstationary components.[8][9][10] A common approach is to apply the Fourier Transform to the original time series to obtain its magnitude and phase spectra. The magnitude spectrum, which determines the power distribution across frequencies, is retained to preserve the global autocorrelation structure. The phase spectrum, which encodes the temporal alignment of frequency components and is often responsible for time-dependent dynamics in the time series (like non-stationarities), is then randomized, typically by replacing it with a set of random phases drawn uniformly from while enforcing conjugate symmetry to ensure a real-valued inverse. Applying the inverse Fourier Transform to the modified spectra yields a strictly stationary surrogate time series:[11] one with the same power spectrum as the original but lacking the temporal structures that caused non-stationarity. This technique is often used in hypothesis tests for probing the stationarity property.[8][10][12][13]