The STATESPACE Procedure

Relation of ARMA and State Space Forms

Every state space model has an ARMA representation, and conversely every ARMA model has a state space representation. This section discusses this equivalence. The following material is adapted from Akaike (1974), where there is a more complete discussion. Pham-Dinh-Tuan (1978) also contains a discussion of this material.

Suppose you are given the following ARMA model:

\[  \bPhi ({B})\mb {x}_{t}=\bTheta ({B})\mb {e}_{t}  \]

or, in more detail,

$\mb {x}_{t}-\bPhi _{1}\mb {x}_{t-1}- {\cdots }-\bPhi _{p}\mb {x}_{t-p} = \mb {e}_{t}+\bTheta _{1}\mb {e}_{t-1}+ {\cdots }+\bTheta _{q}\mb {e}_{t-q}$

(1)

where ${\mb {e}_{t}}$ is a sequence of independent multivariate normal random vectors with mean 0 and variance matrix ${\bSigma _{\mb {ee}}}$, B is the backshift operator (${{B}\mb {x}_{t}=\mb {x}_{t-1}}$), ${\bPhi ({B})}$ and ${\bTheta ({B})}$ are matrix polynomials in B, and ${\mb {x}_{t}}$ is the observed process.

If the roots of the determinantial equation ${{|\bPhi ({B})|} = 0}$ are outside the unit circle in the complex plane, the model can also be written as

\[  \mb {x}_{t} = \bPhi ^{-1}({B})\bTheta ({B})\mb {e}_{t} = \sum _{i=0}^{{\infty }}{\bPsi _{i}\mb {e}_{t-i} }  \]

The ${\bPsi _{i}}$ matrices are known as the impulse response matrices and can be computed as ${\bPhi ^{-1}({B})\bTheta ({B})}$.

You can assume ${p>q}$ since, if this is not initially true, you can add more terms ${\bPhi _{i}}$ that are identically 0 without changing the model.

To write this set of equations in a state space form, proceed as follows. Let ${\mb {x}_{t+i|t}}$ be the conditional expectation of ${\mb {x}_{t+i}}$ given ${\mb {x}_{w}}$ for ${w {\leq } t}$. The following relations hold:

\[  \mb {x}_{t+i|t} = \sum _{j=i}^{{\infty }}{\bPsi _{j}\mb {e}_{t+i-j} }  \]
\[  \mb {x}_{t+i|t+1} = \mb {x}_{t+i|t}+\bPsi _{i-1}\mb {e}_{t+1}  \]

However, from equation (1) you can derive the following relationship:

$\mb {x}_{t+p|t} = \bPhi _{1}\mb {x}_{t+p-1|t}+{\cdots }+\bPhi _{p}\mb {x}_{t}$

(2)

Hence, when ${i=p}$, you can substitute for ${\mb {x}_{t+p|t}}$ in the right-hand side of equation (2) and close the system of equations.

This substitution results in the following model in the state space form ${\mb {z}_{t+1}=\mb {Fz}_{t}+\mb {Ge}_{t+1}}$:

\[  \left[\begin{matrix}  \mb {x}_{t+1}   \\ \mb {x}_{t+2|t+1}   \\ {\vdots }   \\ \mb {x}_{t+p|t+1}   \end{matrix} \right] = \left[\begin{matrix}  0   &  \mb {I}   &  0   &  {\cdots }   &  0   \\ 0   &  0   &  \mb {I}   &  {\cdots }   &  0   \\ {\vdots }   &  {\vdots }   &  {\vdots }   & &  {\vdots }   \\ \bPhi _{p}   &  \bPhi _{p-1}   & & {\cdots }   &  \bPhi _{1}   \\ \end{matrix} \right] \left[\begin{matrix}  \mb {x}_{t}   \\ \mb {x}_{t+1|t}   \\ {\vdots }   \\ \mb {x}_{t+p-1|t}   \end{matrix} \right] + \left[\begin{matrix}  \mb {I}   \\ \bPsi _{1}   \\ {\vdots }   \\ \bPsi _{p-1}   \\ \end{matrix} \right] \mb {e}_{t+1}  \]

Note that the state vector ${\mb {z}_{t}}$ is composed of conditional expectations of ${\mb {x}_{t}}$ and the first r components of ${\mb {z}_{t}}$ are equal to ${\mb {x}_{t}}$.

The state space form can be cast into an ARMA form by solving the system of difference equations for the first r components.

When converting from an ARMA form to a state space form, you can generate a state vector larger than needed; that is, the state space model might not be a minimal representation. When going from a state space form to an ARMA form, you can have nontrivial common factors in the autoregressive and moving average operators that yield an ARMA model larger than necessary.

If the state space form used is not a minimal representation, some but not all components of ${\mb {x}_{t+i|t}}$ might be linearly dependent. This situation corresponds to ${[\bPhi _{p} \bTheta _{p-1} ]}$ being of less than full rank when ${\bPhi ({B})}$ and ${\bTheta ({B})}$ have no common nontrivial left factors. In this case, ${\mb {z}_{t}}$ consists of a subset of the possible components of ${[\mb {x}_{t+i|t}] \quad i=1, 2,{\cdots }, p-1.}$ However, once a component of ${\mb {x}_{t+i|t}}$ (for example, the jth one) is linearly dependent on the previous conditional expectations, then all subsequent jth components of ${\mb {x}_{t+k|t}}$ for ${k > i}$ must also be linearly dependent. Note that in this case, equivalent but seemingly different structures can arise if the order of the components within ${\mb {x}_{t}}$ is changed.