Wold's theorem

From Wikipedia, the free encyclopedia

In statistics, Wold's theorem or Wold representation theorem, named after Herman Wold, says that every covariance-stationary time series (Yt)t can be written as an infinite moving average (MA(\infty)) process of its innovation process. This is also known as the Wold decomposition theorem.

Formally

Y_{t}=\sum_{j=1}^{\infty}b_{j}\varepsilon_{t-j}+\varepsilon_{t}+\eta_{t},

where:

Yt is the considered time series,

\varepsilon_{t} is the innovation proces to the process Yt - it may also be interpreted as the forecast error in prediction applications (actual minus predicted value),

b is the infinite vector of moving average weights (coefficients or parameters), which is absolutely summable \sum_{j=0}^{\infty}|b_{j}| < \infty, and with b0 = 1,

and ηt is a deterministic component, which is zero in the absence of trends in (Yt)t.


The usefulness of the Wold Theorem is that it allows to approximate the dynamic evolution of a variable Yt by a linear model. If the innovations \varepsilon_{t} are independent, then the linear model is the only possible representation. However, when \varepsilon_{t} is merely an uncorrelated but not independent sequence, then the linear model exists but it is not the only representation of the dynamic dependence of the series. In this latter case, it is possible that the linear model is not very useful, and we may have a nonlinear model relating the observed value of Yt with its past evolution. The Wold representation depends on an infinite number of parameters and, consequently, it is not directly useful in practice. To solve this problem, it is approximated using models that have a finite number of parameters, possibly adding an autoregressive part to the equation.