In mathematics and statistics, a stationary process (or a strict/strictly stationary process or strong/strongly stationary process) is a stochastic process whose unconditional joint probability distribution does not change when shifted in time. Consequently, parameters such as mean and variance also do not change over time.
Since stationarity is an assumption underlying many statistical procedures used in time series analysis, non-stationary data are often transformed to become stationary. The most common cause of violation of stationarity is a trend in the mean, which can be due either to the presence of a unit root or of a deterministic trend. In the former case of a unit root, stochastic shocks have permanent effects, and the process is not mean-reverting. In the latter case of a deterministic trend, the process is called a trend-stationary process, and stochastic shocks have only transitory effects after which the variable tends toward a deterministically evolving (non-constant) mean.
A trend stationary process is not strictly stationary, but can easily be transformed into a stationary process by removing the underlying trend, which is solely a function of time. Similarly, processes with one or more unit roots can be made stationary through differencing. An important type of non-stationary process that does not include a trend-like behavior is a cyclostationary process, which is a stochastic process that varies cyclically with time.
For many applications strict-sense stationarity is too restrictive. Other forms of stationarity such as wide-sense stationarity or N-th-order stationarity are then employed. The definitions for different kinds of stationarity are not consistent among different authors (see Other terminology).
Formally, let
\left\{Xt\right\}
FX
(x | |
t1+\tau |
,\ldots,
x | |
tn+\tau |
)
\left\{Xt\right\}
t1+\tau,\ldots,tn+\tau
\left\{Xt\right\}
Since
\tau
FX( ⋅ )
FX
White noise is the simplest example of a stationary process.
An example of a discrete-time stationary process where the sample space is also discrete (so that the random variable may take one of N possible values) is a Bernoulli scheme. Other examples of a discrete-time stationary process with continuous sample space include some autoregressive and moving average processes which are both subsets of the autoregressive moving average model. Models with a non-trivial autoregressive component may be either stationary or non-stationary, depending on the parameter values, and important non-stationary special cases are where unit roots exist in the model.
Let
Y
\left\{Xt\right\}
Xt=Y forallt.
\left\{Xt\right\}
Y
Y
The time average of
Xt
As a further example of a stationary process for which any single realisation has an apparently noise-free structure, let
Y
[0,2\pi]
\left\{Xt\right\}
Xt=\cos(t+Y) fort\inR.
\left\{Xt\right\}
(t+Y)
2\pi
Y
t
Keep in mind that a weakly white noise is not necessarily strictly stationary. Let
\omega
(0,2\pi)
\left\{zt\right\}
zt=\cos(t\omega) (t=1,2,...)
Then
\begin{align} E(zt)&=
1 | |
2\pi |
2\pi | |
\int | |
0 |
\cos(t\omega)d\omega=0,\\ \operatorname{Var}(zt)&=
1 | |
2\pi |
2\pi | |
\int | |
0 |
\cos2(t\omega)d\omega=1/2,\\ \operatorname{Cov}(zt,zj)&=
1 | |
2\pi |
2\pi | |
\int | |
0 |
\cos(t\omega)\cos(j\omega)d\omega=0 \forallt ≠ j. \end{align}
\{zt\}
In, the distribution of
n
n
n
N
\left\{Xt\right\}
A weaker form of stationarity commonly employed in signal processing is known as weak-sense stationarity, wide-sense stationarity (WSS), or covariance stationarity. WSS random processes only require that 1st moment (i.e. the mean) and autocovariance do not vary with respect to time and that the 2nd moment is finite for all times. Any strictly stationary process which has a finite mean and covariance is also WSS.[2]
\left\{Xt\right\}
mX(t)\triangleq\operatornameE[Xt]
KXX(t1,t2)\triangleq\operatorname
E[(X | |
t1 |
-mX(t1))(X
t2 |
-mX(t2))]
The first property implies that the mean function
mX(t)
t1
t2
KXX(t1-t2,0)
the notation is often abbreviated by the substitution
\tau=t1-t2
KXX(\tau)\triangleqKXX(t1-t2,0)
This also implies that the autocorrelation depends only on
\tau=t1-t2
RX(t1,t2)=RX(t1-t2,0)\triangleqRX(\tau).
The third property says that the second moments must be finite for any time
t
The main advantage of wide-sense stationarity is that it places the time-series in the context of Hilbert spaces. Let H be the Hilbert space generated by (that is, the closure of the set of all linear combinations of these random variables in the Hilbert space of all square-integrable random variables on the given probability space). By the positive definiteness of the autocovariance function, it follows from Bochner's theorem that there exists a positive measure
\mu
\omega\xi
t
Xt=\inte-d\omegaλ,
where the integral on the right-hand side is interpreted in a suitable (Riemann) sense. The same result holds for a discrete-time stationary process, with the spectral measure now defined on the unit circle.
When processing WSS random signals with linear, time-invariant (LTI) filters, it is helpful to think of the correlation function as a linear operator. Since it is a circulant operator (depends only on the difference between the two arguments), its eigenfunctions are the Fourier complex exponentials. Additionally, since the eigenfunctions of LTI operators are also complex exponentials, LTI processing of WSS random signals is highly tractable—all computations can be performed in the frequency domain. Thus, the WSS assumption is widely employed in signal processing algorithms.
In the case where
\left\{Xt\right\}
KXX(t1,t2)=\operatorname
E[(X | |
t1 |
-mX(t1))\overline{(X
t2 |
-mX(t2))}]
JXX(t1,t2)=\operatorname
E[(X | |
t1 |
-mX(t1))(X
t2 |
-mX(t2))]
\left\{Xt\right\}
The concept of stationarity may be extended to two stochastic processes.
Two stochastic processes
\left\{Xt\right\}
\left\{Yt\right\}
FXY
(x | |
t1 |
,\ldots,
x | |
tm |
,y | |||||||
|
,\ldots,
y | |||||||
|
)
Two random processes
\left\{Xt\right\}
\left\{Yt\right\}
Two stochastic processes
\left\{Xt\right\}
\left\{Yt\right\}
KXY(t1,t2)=\operatorname
E[(X | |
t1 |
-mX(t1))(Y
t2 |
-mY(t2))]
\tau=t1-t2
N=2
The terminology used for types of stationarity other than strict stationarity can be rather mixed. Some examples follow.
One way to make some time series stationary is to compute the differences between consecutive observations. This is known as differencing. Differencing can help stabilize the mean of a time series by removing changes in the level of a time series, and so eliminating trends. This can also remove seasonality, if differences are taken appropriately (e.g. differencing observations 1 year apart to remove a yearly trend).
Transformations such as logarithms can help to stabilize the variance of a time series.
One of the ways for identifying non-stationary times series is the ACF plot. Sometimes, patterns will be more visible in the ACF plot than in the original time series; however, this is not always the case.[6]
Another approach to identifying non-stationarity is to look at the Laplace transform of a series, which will identify both exponential trends and sinusoidal seasonality (complex exponential trends). Related techniques from signal analysis such as the wavelet transform and Fourier transform may also be helpful.