Every state space model has an ARMA representation, and conversely every ARMA model has a state space representation. This section discusses this equivalence. The following material is adapted from Akaike (1974), where there is a more complete discussion. Pham (1978) also contains a discussion of this material.
Suppose you are given the following ARMA model:
or, in more detail,
| 
                            | 
                          
                            
                            
                     
                         (1)  | 
                          
                          
                  
where 
 is a sequence of independent multivariate normal random vectors with mean 0 and variance matrix 
, B is the backshift operator (
), 
 and 
 are matrix polynomials in B, and 
 is the observed process. 
            
If the roots of the determinantial equation 
 are outside the unit circle in the complex plane, the model can also be written as 
            
The 
 matrices are known as the impulse response matrices and can be computed as 
. 
You can assume 
 since, if this is not initially true, you can add more terms 
 that are identically 0 without changing the model. 
            
To write this set of equations in a state space form, proceed as follows. Let 
 be the conditional expectation of 
 given 
 for 
. The following relations hold: 
            
However, from equation (1) you can derive the following relationship:
| 
                            | 
                          
                            
                            
                     
                         (2)  | 
                          
                          
                  
Hence, when 
, you can substitute for 
 in the right-hand side of equation (2) and close the system of equations. 
            
This substitution results in the following model in the state space form 
: 
            
![\[  \left[\begin{matrix}  \mb{x}_{t+1}   \\ \mb{x}_{t+2|t+1}   \\ {\vdots }   \\ \mb{x}_{t+p|t+1}   \end{matrix} \right] = \left[\begin{matrix}  0   &  \mb{I}   &  0   &  {\cdots }   &  0   \\ 0   &  0   &  \mb{I}   &  {\cdots }   &  0   \\ {\vdots }   &  {\vdots }   &  {\vdots }   & &  {\vdots }   \\ \bPhi _{p}   &  \bPhi _{p-1}   & & {\cdots }   &  \bPhi _{1}   \\ \end{matrix} \right] \left[\begin{matrix}  \mb{x}_{t}   \\ \mb{x}_{t+1|t}   \\ {\vdots }   \\ \mb{x}_{t+p-1|t}   \end{matrix} \right] + \left[\begin{matrix}  \mb{I}   \\ \bPsi _{1}   \\ {\vdots }   \\ \bPsi _{p-1}   \\ \end{matrix} \right] \mb{e}_{t+1}  \]](images/etsug_statespa0205.png)
Note that the state vector 
 is composed of conditional expectations of 
 and the first r components of 
 are equal to 
. 
            
The state space form can be cast into an ARMA form by solving the system of difference equations for the first r components.
When converting from an ARMA form to a state space form, you can generate a state vector larger than needed; that is, the state space model might not be a minimal representation. When going from a state space form to an ARMA form, you can have nontrivial common factors in the autoregressive and moving average operators that yield an ARMA model larger than necessary.
If the state space form used is not a minimal representation, some but not all components of 
 might be linearly dependent. This situation corresponds to 
 being of less than full rank when 
 and 
 have no common nontrivial left factors. In this case, 
 consists of a subset of the possible components of 
 However, once a component of 
 (for example, the jth one) is linearly dependent on the previous conditional expectations, then all subsequent jth components of 
 for 
 must also be linearly dependent. Note that in this case, equivalent but seemingly different structures can arise if the order
               of the components within 
 is changed.