|The LOGISTIC Procedure|
|Conditional Logistic Regression|
The method of maximum likelihood described in the preceding sections relies on large-sample asymptotic normality for the validity of estimates and especially of their standard errors. When you do not have a large sample size compared to the number of parameters, this approach might be inappropriate and might result in biased inferences. This situation typically arises when your data are stratified and you fit intercepts to each stratum so that the number of parameters is of the same order as the sample size. For example, in a matched pairs study with pairs and covariates, you would estimate intercept parameters and slope parameters. Taking the stratification into account by "conditioning out" (and not estimating) the stratum-specific intercepts gives consistent and asymptotically normal MLEs for the slope coefficients. See Breslow and Day (1980) and Stokes, Davis, and Koch (2000) for more information. If your nuisance parameters are not just stratum-specific intercepts, you can perform an exact conditional logistic regression .
For each stratum , , number the observations as so that indexes the th observation in the th stratum. Denote the covariates for observation as and its binary response as , and let , , and . Let the dummy variables , be indicator functions for the strata ( if the observation is in stratum ), and denote for observation , , and . Denote ) and . Arrange the observations in each stratum so that for , and for . Suppose all observations have unit frequency.
Consider the binary logistic regression model written as
where the parameter vector consists of , is the intercept for stratum , and is the parameter vector for the covariates.
From the section Determining Observations for Likelihood Contributions, you can write the likelihood contribution of observation as
where when the response takes Ordered Value 1, and otherwise.
The full likelihood is
Unconditional likelihood inference is based on maximizing this likelihood function.
When your nuisance parameters are the stratum-specific intercepts , and the slopes are your parameters of interest, "conditioning out" the nuisance parameters produces the conditional likelihood (Lachin; 2000)
where the summation is over all subsets of observations chosen from the observations in stratum . Note that the nuisance parameters have been factored out of this equation.
For conditional asymptotic inference, maximum likelihood estimates of the regression parameters are obtained by maximizing the conditional likelihood, and asymptotic results are applied to the conditional likelihood function and the maximum likelihood estimators. A relatively fast method of computing this conditional likelihood and its derivatives is given by Gail, Lubin, and Rubinstein (1981) and Howard (1972). The default optimization techniques, which are the same as those implemented by the NLP procedure in SAS/OR software, are as follows:
Newton-Raphson with ridging when the number of parameters
conjugate gradient when
Sometimes the log likelihood converges but the estimates diverge. This condition is flagged by having inordinately large standard errors for some of your parameter estimates, and can be monitored by specifying the ITPRINT option. Unfortunately, broad existence criteria such as those discussed in the section Existence of Maximum Likelihood Estimates do not exist for this model. It might be possible to circumvent such a problem by standardizing your independent variables before fitting the model.
Diagnostics are used to indicate observations that might have undue influence on the model fit or that might be outliers. Further investigation should be performed before removing such an observation from the data set.
The derivations in this section use an augmentation method described by Storer and Crowley (1985), which provides an estimate of the "one-step" DFBETAS estimates advocated by Pregibon (1984). The method also provides estimates of conditional stratum-specific predicted values, residuals, and leverage for each observation.
Following Storer and Crowley (1985), the log-likelihood contribution can be written as
and the subscript on matrices indicates the submatrix for the stratum, , and . Then the gradient and information matrix are
and where is the conditional stratum-specific probability that subject in stratum is a case, the summation on is over all subsets from of size that contain the index , and the summation on is over all subsets from of size that contain the indices and .
To produce the true one-step estimate , start at the MLE , delete the th observation, and use this reduced data set to compute the next Newton-Raphson step. Note that if there is only one event or one nonevent in a stratum, deletion of that single observation is equivalent to deletion of the entire stratum. The augmentation method does not take this into account.
The augmented model is
where has a in the th coordinate, and use as the initial estimate for . The gradient and information matrix before the step are
Inserting the and into the Gail, Lubin, and Rubinstein (1981) algorithm provides the appropriate estimates of and . Indicate these estimates with , , , and .
DFBETA is computed from the information matrix as
For each observation in the data set, a DFBETA statistic is computed for each parameter , , and standardized by the standard error of from the full data set to produce the estimate of DFBETAS.
The estimated residuals are obtained from , and the weights, or predicted probabilities, are then . The residuals are standardized and reported as (estimated) Pearson residuals:
where is the number of events in the observation and is the number of trials.
The estimated leverage is defined as
This definition of leverage produces different values from those defined by Pregibon (1984), Moolgavkar, Lustbader, and Venzon (1985), and Hosmer and Lemeshow (2000); however, it has the advantage that no extra computations beyond those for the DFBETAS are required.
For events/trials MODEL syntax, treat each observation as two observations (the first for the nonevents and the second for the events) with frequencies and , and augment the model with a matrix instead of a single vector. Writing in the preceding section results in the following gradient and information matrix.
The predicted probabilities are then , while the leverage and the DFBETAS are produced from in a fashion similar to that for the preceding single-trial equations.