The PANEL Procedure

One-Way Random-Effects Model

The specification for the one-way random-effects model is

\[  u_{it}={\nu }_{i}+{\epsilon }_{it}  \]

Let $\mb{Z} _{0}=\mr{diag}(\mb{J} _{T_{i}}$), ${\mb{P} _{0}=\mr{diag}({\bar{\mb{J}}}_{T_{i}})}$, and $\mb{Q} _{0}=\mr{diag}(\mb{E} _{T_{i}})$, with ${\bar{\mb{J}}}_{T_{i}}=\mb{J} _{T_{i}}/\mi{T} _{i}$ and ${\mb{E} _{T_{i}}=\mb{I} _{T_{i}}-{\bar{\mb{J}}}_{T_{i}} }$. Define ${\tilde{\mb{X} }_{s}=\mb{Q} _{0}\mb{X} _{s} }$ and ${\tilde{\mb{y} }=\mb{Q} _{0}\mb{y} }$ and $\mb{J}$ as a vector of ones ${T_{i}}$ long.

In the one-way model, estimation proceeds in a two-step fashion. First, you obtain estimates of the variance of the ${ {\sigma }^{2}_{{\epsilon } } }$ and ${{\sigma }^2_{{\nu }} }$. There are multiple ways to derive these estimates; PROC PANEL provides four options. All four options are valid for balanced or unbalanced panels. Once these estimates are in hand, they are used to form a weighting factor $\theta $, and estimation proceeds via OLS on partial deviations from group means.

PROC PANEL gives the following options for variance component estimators.

Fuller and Battese’s Method

The Fuller and Battese method for estimating variance components can be obtained with the option VCOMP = FB and the option RANONE. The variance components are given by the following equations (For the approach in the two-way model, see Baltagi and Chang (1994); Fuller and Battese (1974)). Let

\[  \mi{R} ({\nu })=\mb{y} ^{'}\mb{Z} _{0}( \mb{Z} ^{'}_{0}\mb{Z} _{0})^{-1} \mb{Z} ^{'}_{0}\mb{y}  \]
\[  \mi{R} ({\beta }|{\nu })=(( \tilde{\mb{X} }^{'}_{s} \tilde{\mb{X} }_{s})^{-1} \tilde{\mb{X} }^{'}_{s}\tilde{\mb{y} } )^{'} ( \tilde{\mb{X} }^{'}_{s}\tilde{\mb{y} })  \]
\[  \mi{R} ({\beta })=(\mb{X}_ s ^{'}\mb{y} )^{'} (\mb{X}_ s ^{'}\mb{X}_ s ) ^{-1}\mb{X}_ s ^{'}\mb{y}  \]
\[  \mi{R} ({\nu }|{\beta })=\mi{R} ({\beta }|{\nu })+\mi{R} ({\nu })-\mi{R} ({\beta })  \]

The estimator of the error variance is given by

\[  \hat{{\sigma }}_{{\epsilon }}^{2}= (\mb{y} ^{'}\mb{y}- \mi{R} ({\beta }|{\nu })- \mi{R} ({\nu }) ) / (\mi{M}-\mi{N}-(\mi{K} -1))  \]

If the NOINT option is specified, the estimator is

\[  \hat{{\sigma }}_{{\epsilon }}^{2}= (\mb{y} ^{'}\mb{y}- \mi{R} ({\beta }|{\nu })- \mi{R} ({\nu }) ) / (\mi{M}-\mi{N}-\mi{K})  \]

The estimator of the cross-sectional variance component is given by

\[  \hat{{\sigma }}_{{\nu }}^{2}= (\mi{R} ({\nu }|{\beta })- (\mi{N} -1)\hat{{\sigma }}_{{\epsilon }}^{2} ) /(\mi{M}- \mr{tr}( \mb{Z} ^{'}_{0}\mb{X}_ s (\mb{X}_ s ^{'}\mb{X}_ s )^{-1}\mb{X}_ s ^{'}\mb{Z} _{0} ))  \]

Note that the error variance is the variance of the residual of the within estimator.

According to Baltagi and Chang (1994), the Fuller and Battese method is appropriate to apply to both balanced and unbalanced data. The Fuller and Battese method is the default for estimation of one-way random-effects models with balanced panels. However, the Fuller and Battese method does not always obtain nonnegative estimates for the cross section (or group) variance. In the case of a negative estimate, a warning is printed and the estimate is set to zero.

Wansbeek and Kapteyn’s Method

The Wansbeek and Kapteyn method for estimating variance components can be obtained by setting VCOMP = WK (together with the option RANONE). The estimation of the one-way unbalanced data model is performed by using a specialization (Baltagi and Chang, 1994) of the approach used by Wansbeek and Kapteyn (1989) for unbalanced two-way models. The Wansbeek and Kapteyn method is the default for unbalanced data. If just RANONE is specified, without the VCOMP= option, PROC PANEL estimates the variance component under Wansbeek and Kapteyn’s method.

The estimation of the variance components is performed by using a quadratic unbiased estimation (QUE) method. This involves focusing on quadratic forms of the centered residuals, equating their expected values to the realized quadratic forms, and solving for the variance components.

Let

\[  q_{1}=\tilde{\mb{u} }^{'}\mb{Q} _{0}\tilde{\mb{u} }  \]
\[  q_{2}=\tilde{\mb{u} }^{'}\mb{P} _{0}\tilde{\mb{u} }  \]

where the residuals ${\tilde{\mb{u}}}$ are given by $\tilde{\mb{u}} = (\mb{I} _{M}-\mb{j} _{M} \mb{j} {’}_{M}/\mi{M} ) (\mb{y}-\mb{X} _{s} (\tilde{\mb{X}} {’}_{s} \tilde{\mb{X}}_{s})^{-1} \tilde{\mb{X}}_{s}{’}\tilde{\mb{y}} $ if there is an intercept and by $\tilde{\mb{u}} = \mb{y}-\mb{X} _{s} ( \tilde{\mb{X}} {’}_{s} \tilde{\mb{X}}_{s})^{-1} \tilde{\mb{X}} {’}_{s}\tilde{\mb{y}}$ if there is not. A vector of M ones is represented by $\mb{j}$.

Consider the expected values

\[  \mi{E} (q_{1})=(\mi{M}-\mi{N}-(\mi{K} -1)) {\sigma }^{2}_{{\epsilon }}  \]
\[  \mi{E} (q_{2})=(\mi{N} -1+ \mr{tr}[( \mb{X} ^{'}_{s}\mb{Q} _{0}\mb{X} _{s})^{-1} \mb{X} ^{'}_{s}\mb{P} _{0}\mb{X} _{s}]- \mr{tr}[( \mb{X} ^{'}_{s}\mb{Q} _{0}\mb{X} _{s})^{-1} \mb{X} ^{'}_{s} {\bar{\mb{J}}}_{M}\mb{X} _{s}] ) {\sigma }^{2}_{{\epsilon }} \]
\[ \; \;  \; \;  [\mi{M}-(\sum _{i} \mi{T} ^{2}_{i}/\mi{M} )] {\sigma }^{2}_{{\nu }} \]

where ${ \hat{{\sigma }}^{2}_{{\epsilon }} }$ and ${ \hat{{\sigma }}^{2}_{{\nu }} }$ are obtained by equating the quadratic forms to their expected values.

The estimator of the error variance is the residual variance of the within estimate. The Wansbeek and Kapteyn method can also generate negative variance components estimates.

Wallace and Hussain’s Method

The Wallace and Hussain method for estimating variance components can be obtained by setting VCOMP = WH (together with the option RANONE). Wallace-Hussain estimates start from OLS residuals on a data that are assumed to exhibit groupwise heteroscedasticity. As in the Wansbeek and Kapteyn method, you start with

\[  q_{1}=\tilde{\mb{u}}_{OLS}^{'}\mb{Q} _{0}\tilde{\mb{u}}_{OLS}  \]
\[  q_{2}=\tilde{\mb{u}}_{OLS}^{'}\mb{P} _{0}\tilde{\mb{u}}_{OLS}  \]

However, instead of using the 'true' errors, you substitute the OLS residuals. You solve the system

\[  \mi{E} (\hat{q}_{1}) = \mi{E} (\hat{u}_{OLS}^{'}\mb{Q} _{0}\hat{u}_{OLS}) = \delta _{11} \hat{{\sigma }}^{2}_{{\nu }} + \delta _{12} \hat{{\sigma }}^{2}_{{\epsilon }}  \]
\[  \mi{E} (\hat{q}_{2}) = \mi{E} (\hat{u}_{OLS}^{'}\mb{P} _{0}\hat{u}_{OLS}) = \delta _{21} \hat{{\sigma }}^{2}_{{\nu }} + \delta _{22} \hat{{\sigma }}^{2}_{{\epsilon }}  \]

The constants $\delta _{11}, \delta _{12}, \delta _{21}, \delta _{22}$ are given by

\[  \delta _{11} = \mr{tr}\left(\left(\mb{X} ^{'}\mb{X} \right)^{-1}\mb{X} ^{'}\mb{Z}_0 \mb{Z}_0 ^{'}\mb{X} \right) - \mr{tr}\left(\left(\mb{X} ^{'}\mb{X} \right)^{-1}\mb{X} ^{'}\mb{P}_0 \mb{X} \left(\mb{X} ^{'}\mb{X} \right)^{-1}\mb{X} ^{'}\mb{Z}_0 \mb{Z}_0 ^{'}\mb{X} \right)  \]
\[  \delta _{12} = M - N - K + \mr{tr}\left(\left(\mb{X} ^{'}\mb{X} \right)^{-1}\mb{X} ^{'}\mb{P}_0 \mb{X} \right)  \]
\[  \delta _{21} = M - 2tr\left(\left(\mb{X} ^{'}\mb{X} \right)^{-1}\mb{X} ^{'}\mb{Z}_0 \mb{Z}_0 ^{'}\mb{X} \right) + \mr{tr}\left(\left(\mb{X} ^{'}\mb{X} \right)^{-1}\mb{X} ^{'}\mb{P}_0 \mb{X} \right)  \]
\[  \delta _{22} = N - \mr{tr}\left(\left(\mb{X} ^{'}\mb{X} \right)^{-1}\mb{X} ^{'}\mb{P}_0 \mb{X} \right)  \]

where tr() is the trace operator on a square matrix.

Solving this system produces the variance components. This method is applicable to balanced and unbalanced panels. However, there is no guarantee of positive variance components. Any negative values are fixed equal to zero.

Nerlove’s Method

The Nerlove method for estimating variance components can be obtained by setting VCOMP = NL. The Nerlove method (see Baltagi 1995, p. 17) is assured to give estimates of the variance components that are always positive. Furthermore, it is simple in contrast to the previous estimators.

If $\gamma _\mi {i} $ is the $\emph{i}$th fixed effect, Nerlove’s method uses the variance of the fixed effects as the estimate of ${ \hat{{\sigma }}^{2}_{{\nu }} }$. You have ${ \hat{{\sigma }}^{2}_{{\nu }} } = \sum _\mi {i = 1} ^\mi {N} \frac{\left(\gamma _\mi {i}- \bar{\gamma }\right)^2}{\mi{N}- 1}$, where $\bar{\gamma }$ is the mean fixed effect. The estimate of ${\sigma }^{2}_{{\epsilon }}$ is simply the residual sum of squares of the one-way fixed-effects regression divided by the number of observations.

Transformation and Estimation

After you calculate the variance components from any method, the next task is to estimate the regression model of interest. For each individual, you form a weight ($\theta _\mi {i} $) as

\[  \theta _\mi {i} = 1 - \sigma _{\epsilon } / w_\mi {i}  \]
\[  w_{i}^{2} = T_{i}{\sigma }^{2}_{{\nu }} + {\sigma }^{2}_{{\epsilon }}  \]

where $T_{i}$ is the $\emph{i}$th cross section’s time observations.

Taking the $\theta _\mi {i} $, you form the partial deviations,

\[  \tilde{y}_\mi {it} = y_\mi {it}- \theta _\mi {i} \bar{y}_\mi {i \cdot }  \]
\[  \tilde{x}_\mi {it} = x_\mi {it}- \theta _\mi {i} \bar{x}_\mi {i \cdot }  \]

where $\bar{y}_\mi {i \cdot }$ and $\bar{x}_\mi {i \cdot }$ are cross-sectional means of the dependent variable and independent variables (including the constant if any), respectively.

The random effects $\beta $ is then the result of simple OLS on the transformed data.