The HPREG procedure supports a variety of fit statistics that you can specify as criteria for the CHOOSE=, SELECT=, and STOP= options in the SELECTION statement. The following statistics are available:
Adjusted R-square statistic (Darlington, 1968; Judge et al., 1985)
Akaike’s information criterion (Akaike, 1969; Judge et al., 1985)
Corrected Akaike’s information criterion (Hurvich and Tsai, 1989)
Schwarz Bayesian information criterion (Schwarz, 1978; Judge et al., 1985)
Predicted residual sum of squares statistic
Significance used to assess an effect’s contribution to the fit when it is added to or removed from a model
Average square error over the validation data
When you use SL as a criterion for effect selection, the definition depends on whether an effect is being considered as a drop or an add candidate. If the current model has p parameters excluding the intercept, and if you denote its residual sum of squares by and you add an effect with k degrees of freedom and denote the residual sum of squares of the resulting model by , then the F statistic for entry with k numerator degrees of freedom and denominator degrees of freedom is given by
where n is number of observations used in the analysis. The significance level for entry is the p-value of this F statistic, and is deemed significant if it is smaller than the SLENTRY limit. Among several such add candidates, the effect with the smallest p-value (most significant) is deemed best.
If you drop an effect with k degrees of freedom and denote the residual sum of squares of the resulting model by , then the F statistic for removal with k numerator degrees of freedom and denominator degrees of freedom is given by
where n is number of observations used in the analysis. The significance level for removal is the p-value of this F statistic, and the effect is deemed not significant if this p-value is larger than the SLSTAY limit. Among several such removal candidates, the effect with the largest p-value (least significant) is deemed the best removal candidate.
It is known that the “F-to-enter” and “F-to-delete” statistics do not follow an F distribution (Draper, Guttman, and Kanemasu, 1971).. Hence the SLENTRY and SLSTAY values cannot reliably be viewed as probabilities. One way to address this difficulty is to replace hypothesis testing as a means of selecting a model with information criteria or out-of-sample prediction criteria. While Harrell (2001) points out that information criteria were developed for comparing only prespecified models, Burnham and Anderson (2002) note that AIC criteria have routinely been used for several decades for performing model selection in time series analysis.
Table 12.5 provides formulas and definitions for these fit statistics.
Table 12.5: Formulas and Definitions for Model Fit Summary Statistics
Statistic |
Definition or Formula |
---|---|
n |
Number of observations |
p |
Number of parameters including the intercept |
|
Estimate of pure error variance from fitting the full model |
|
Total sum of squares corrected for the mean for the |
|
Error sum of squares |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|