The NORMAL option in the FIT statement performs multivariate and univariate tests of normality.
The three multivariate tests provided are Mardia’s skewness test and kurtosis test (Mardia 1970) and the Henze-Zirkler
test (Henze and Zirkler 1990). The two univariate tests provided are the Shapiro-Wilk W test and the Kolmogorov-Smirnov test. (For details on the univariate
tests, refer to "Goodness-of-Fit Tests" section in "The UNIVARIATE Procedure" chapter in the Base SAS Procedures Guide.) The null hypothesis for all these tests is that the residuals are normally distributed.
For a random sample
,
, where d is the dimension of
and n is the number of observations, a measure of multivariate skewness is
![\[ b_{1,d} = \frac{1}{n^{2}} \sum _{i=1}^{n} \sum _{j=1}^{n}{[ ( X_{i} - {\mu })’ {\bS }^{-1} (X_{j} - {\mu })]^{3} } \]](images/etsug_model0299.png)
where S is the sample covariance matrix of X. For weighted regression, both S and
are computed by using the weights supplied by the WEIGHT statement or the _WEIGHT_ variable.
Mardia showed that under the null hypothesis
is asymptotically distributed as
. For small samples, Mardia’s skewness test statistic is calculated with a small sample correction formula, given by
where the correction factor k is given by
. Mardia’s skewness test statistic in PROC MODEL uses this small sample corrected formula.
A measure of multivariate kurtosis is given by
![\[ b_{2,d} = \frac{1}{n} \sum _{i=1}^{n}{[( X_{i} - {\mu })^{'} {\bS }^{-1} ( X_{i} - {\mu })]^{2} } \]](images/etsug_model0305.png)
Mardia showed that under the null hypothesis,
is asymptotically normally distributed with mean
and variance
.
The Henze-Zirkler test is based on a nonnegative functional
that measures the distance between two distribution functions and has the property that
![\[ D(\mr{N}_{d}(0, I_{d}), Q) = 0 \]](images/etsug_model0310.png)
if and only if
![\[ Q = \mr{N}_{d}(0, I_{d}) \]](images/etsug_model0311.png)
where
is a d-dimensional normal distribution.
The distance measure
can be written as
![\[ D_{{\beta }}( P, Q ) = \int _{\mr{R}^{d}}^{}{| \hat{P}(t) - \hat{Q}(t) |^{2} {\varphi }_{{\beta }}(t) dt} \]](images/etsug_model0313.png)
where
and
are the Fourier transforms of P and Q, and
is a weight or a kernel function. The density of the normal distribution
is used as
![\[ {\varphi }_{{\beta }}(t) = ( 2{\pi }{\beta }^{2})^{\frac{-d}{2}} \mr{exp} ( \frac{- |t|^{2}}{2{\beta }^{2}} ), t ~ {\in }~ \mr{R}^{d} \]](images/etsug_model0318.png)
where
.
The parameter
depends on
as
![\[ {\beta }_{d}(n) = \frac{1}{\sqrt {2}}( \frac{2d+1}{4} )^{1/(d+4)} n^{1/(d+4)} \]](images/etsug_model0322.png)
The test statistic computed is called
and is approximately distributed as a lognormal. The lognormal distribution is used to compute the null hypothesis probability.

where
![\[ |Y_{j} - Y_{k}|^{2} = (X_{j} - X_{k})’ {\bS }^{-1} (X_{j} - X_{k}) \]](images/etsug_model0325.png)
![\[ |Y_{j}|^{2} = (X_{j} - \bar{X})’ {\bS }^{-1} (X_{j} - \bar{X}) \]](images/etsug_model0326.png)
Monte Carlo simulations suggest that
has good power against distributions with heavy tails.
The Shapiro-Wilk W test is computed only when the number of observations (n ) is less than 2000 while computation of the Kolmogorov-Smirnov test statistic requires at least 2000 observations.
The following is an example of the output produced by the NORMAL option.
proc model data=test2; y1 = a1 * x2 * x2 - exp( d1*x1); y2 = a2 * x1 * x1 + b2 * exp( d2*x2); fit y1 y2 / normal ; run;
Figure 26.40: Normality Test Output