The REG Procedure

Criteria Used in Model-Selection Methods

When many significance tests are performed, each at a level of, for example, 5%, the overall probability of rejecting at least one true null hypothesis is much larger than 5%. If you want to guard against including any variables that do not contribute to the predictive power of the model in the population, you should specify a very small SLE= significance level for the FORWARD and STEPWISE methods and a very small SLS= significance level for the BACKWARD and STEPWISE methods.

In most applications, many of the variables considered have some predictive power, however small. If you want to choose the model that provides the best prediction computed using the sample estimates, you need only to guard against estimating more parameters than can be reliably estimated with the given sample size, so you should use a moderate significance level, perhaps in the range of 10% to 25%.

In addition to R square, the $C_ p$ statistic is displayed for each model generated in the model-selection methods. The $C_ p$ statistic is proposed by Mallows (1973) as a criterion for selecting a model. It is a measure of total squared error defined as

\[ C_ p = \frac{\mbox{SSE}_ p}{s^2} - (N - 2p) \]

where $s^2$ is the MSE for the full model, and $\mbox{SSE}_ p$ is the sum-of-squares error for a model with p parameters including the intercept, if any. If $C_ p$ is plotted against p, Mallows recommends the model where $C_ p$ first approaches p. When the right model is chosen, the parameter estimates are unbiased, and this is reflected in $C_ p$ near p. For further discussion, see Daniel and Wood (1980).

The adjusted R square statistic is an alternative to R square that is adjusted for the number of parameters in the model. The adjusted R square statistic is calculated as

\[ \mbox{ADJRSQ} = 1 - \frac{(n - i)(1 - R^2)}{n - p} \]

where n is the number of observations used in fitting the model, and i is an indicator variable that is 1 if the model includes an intercept, and 0 otherwise.