Displayed Output

Many of the more specialized tables are described in detail in previous sections. Most of the formulas for the statistics are in Chapter 4, Introduction to Regression Procedures, while other formulas can be found in the section Model Fit and Diagnostic Statistics and the section Influence Statistics.

The analysis-of-variance table includes the following:

  • the Source of the variation, Model for the fitted regression, Error for the residual error, and C Total for the total variation after correcting for the mean. The Uncorrected Total Variation is produced when the NOINT option is used.

  • the degrees of freedom (DF) associated with the source

  • the Sum of Squares for the term

  • the Mean Square, the sum of squares divided by the degrees of freedom

  • the F Value for testing the hypothesis that all parameters are zero except for the intercept. This is formed by dividing the mean square for Model by the mean square for Error.

  • the Prob>F, the probability of getting a greater statistic than that observed if the hypothesis is true. This is the significance probability.

Other statistics displayed include the following:

  • Root MSE is an estimate of the standard deviation of the error term. It is calculated as the square root of the mean square error.

  • Dep Mean is the sample mean of the dependent variable.

  • C.V. is the coefficient of variation, computed as 100 times Root MSE divided by Dep Mean. This expresses the variation in unitless values.

  • R-square is a measure between 0 and 1 that indicates the portion of the (corrected) total variation that is attributed to the fit rather than left to residual error. It is calculated as SS(Model) divided by SS(Total). It is also called the coefficient of determination. It is the square of the multiple correlation—in other words, the square of the correlation between the dependent variable and the predicted values.

  • Adj R-square, the adjusted , is a version of that has been adjusted for degrees of freedom. It is calculated as


    where is equal to 1 if there is an intercept and 0 otherwise, is the number of observations used to fit the model, and is the number of parameters in the model.

The parameter estimates and associated statistics are then displayed, and they include the following:

  • the Variable used as the regressor, including the name Intercept to represent the estimate of the intercept parameter

  • the degrees of freedom (DF) for the variable. There is one degree of freedom unless the model is not full rank.

  • the Parameter Estimate

  • the Standard Error, the estimate of the standard deviation of the parameter estimate

  • T for H0: Parameter=0, the test that the parameter is zero. This is computed as the Parameter Estimate divided by the Standard Error.

  • the Prob > |T|, the probability that a statistic would obtain a greater absolute value than that observed given that the true parameter is zero. This is the two-tailed significance probability.

If model-selection methods other than NONE, RSQUARE, ADJRSQ, and CP are used, the analysis-of-variance table and the parameter estimates with associated statistics are produced at each step. Also displayed are the following:

  • C(p), which is Mallows’ statistic

  • bounds on the condition number of the correlation matrix for the variables in the model (Berk 1977)

After statistics for the final model are produced, the following is displayed when the method chosen is FORWARD, BACKWARD, or STEPWISE:

  • a Summary table listing Step number, Variable Entered or Removed, Partial and Model R-square, and C(p) and F statistics

The RSQUARE method displays its results beginning with the model containing the fewest independent variables and producing the largest . Results for other models with the same number of variables are then shown in order of decreasing , and so on, for models with larger numbers of variables. The ADJRSQ and CP methods group models of all sizes together and display results beginning with the model having the optimal value of adjusted and , respectively.

For each model considered, the RSQUARE, ADJRSQ, and CP methods display the following:

  • Number in Model or IN, the number of independent variables used in each model

  • R-square or RSQ, the squared multiple correlation coefficient

If the B option is specified, the RSQUARE, ADJRSQ, and CP methods produce the following:

  • Parameter Estimates, the estimated regression coefficients

If the B option is not specified, the RSQUARE, ADJRSQ, and CP methods display the following:

  • Variables in Model, the names of the independent variables included in the model