The INTPOINT Procedure |
The simplex algorithm, developed shortly after World War II, was for many years the main method used to solve linear programming problems. Over the last fifteen years, however, the interior point algorithm has been developed. This algorithm also solves linear programming problems. From the start it showed great theoretical promise, and considerable research in the area resulted in practical implementations that performed competitively with the simplex algorithm. More recently, interior point algorithms have evolved to become superior to the simplex algorithm, in general, especially when the problems are large.
There are many variations of interior point algorithms. PROC INTPOINT uses the Primal-Dual with Predictor-Corrector algorithm. More information on this particular algorithm and related theory can be found in the texts by Roos, Terlaky, and Vial (1997), Wright (1996), and Ye (1996).
There exists an equivalent problem, the dual problem, stated as
The interior point algorithm solves the system of equations to satisfy the Karush-Kuhn-Tucker (KKT) conditions for optimality:
These are the conditions for feasibility, with the
complementarity
condition added. Complementarity forces the optimal
objectives of the primal and dual to be equal,
, as
Before the optimum is reached, a solution may not satisfy the KKT conditions:
The interior point algorithm works by using Newton's method to find a
direction
to move
from the current solution
toward a better solution:
where is the step length and is assigned a value as large as possible but not so
large that an
or
is "too close" to zero.
The direction in which to move is found using
To greatly improve performance, the third equation is changed to
The effect now is to find a direction in which to move to reduce
infeasibilities and to reduce the complementarity toward zero,
but if any
is too close to zero, it is "nudged out" to
,
and any
that is larger than
is "nudged into"
.
A
close to or equal to 0.0 biases a direction toward the
optimum,
and a value of
close to or equal to 1.0 "centers" the
direction toward a point where all pairwise products
.
Such points make up the central path in the interior.
Although centering directions make little, if any, progress in reducing
and moving the solution closer to the optimum,
substantial progress toward the optimum can usually be made in
the next iteration.
The central path is crucial to why the interior point algorithm is so
efficient.
As is decreased, this path "guides"
the algorithm to the optimum through the interior of feasible space.
Without centering, the algorithm would find a series of solutions near each other
close to the boundary of feasible space.
Step lengths along the direction
would be small and many
more iterations would probably be required to reach the optimum.
That in a nutshell is the primal-dual interior point algorithm.
Varieties of the algorithm differ in the way and
are chosen
and the direction adjusted during each iteration.
A wealth of information can be found in the texts by
Roos, Terlaky, and Vial (1997), Wright (1996), and Ye (1996).
The calculation of the direction is the most time-consuming step
of the interior point algorithm.
Assume the th iteration is being performed,
so the subscript and superscript
can be dropped from the algebra:
Rearranging the second equation,
Rearranging the third equation,
Equating these two expressions for and rearranging,
Substituting into the first direction equation,
The fact that the nonzeros in have a constant pattern is exploited by all
interior point algorithms and is a major reason for their
excellent performance.
Before iterations begin,
is examined and its rows and
columns are symmetrically permuted so that during Cholesky
factorization, the number of
fill-ins created is smaller.
A list of arithmetic operations to perform the factorization is saved
in concise computer data structures (working with
memory locations rather than actual numerical values).
This is called symbolic factorization.
During iterations, when memory has been initialized with numerical values,
the operations list is performed sequentially.
Determining how the factorization should be performed again and again
is unnecessary.
Complementarity is calculated at
and compared with the complementarity at the starting point
, and the success of the affine step is gauged.
If the affine step was successful in reducing the complementarity
by a substantial amount, the need for centering is not great,
and
in the following linear system
is assigned a value close to zero.
If, however, the affine step was unsuccessful,
centering would be beneficial,
and
in the following linear system
is assigned a value closer to 1.0.
The value of
is therefore adaptively altered depending on the
progress made toward the optimum.
A second linear system is solved to determine a centering vector
from
:
Although the Predictor-Corrector variant entails solving two linear
systems instead of one, fewer iterations are usually required to reach
the optimum.
The additional overhead of calculating the second linear system
is small, as the factorization of the
matrix has already been
performed to solve the first linear system.
If the LP had upper bounds ( where
is
the upper bound vector), then the primal and dual problems,
the duality gap, and the KKT conditions would have to be expanded.
The primal linear program to be solved is
where is split into
and
. Let
be
primal slack so that
, and associate dual variables
with these constraints.
The interior point algorithm solves the system of equations to
satisfy the Karush-Kuhn-Tucker (KKT) conditions for
optimality:
These are the conditions for feasibility, with the
complementarity
conditions and
added.
Complementarity forces the optimal objectives of the primal and dual
to be equal,
, as
Before the optimum is reached, a solution might not satisfy the KKT conditions:
The calculations of the interior point algorithm can easily be derived in a fashion similar to calculations for when an LP has no upper bounds. See the paper by Lustig, Marsten, and Shanno (1992).
In some iteration , the affine step system that must be
solved is
Therefore, the computations involved in solving the affine step are
A second linear system is solved to determine a centering vector
from
:
Therefore, the computations involved in solving the centering step are
The algebra in this section has been simplified by assuming that all
variables have finite upper bounds.
If the number of variables with finite upper bounds , you need to change the
algebra to reflect
that the
and
matrices have dimension
or
.
Other computations need slight modification.
For example, the average complementarity is
An important point is that any upper bounds can be handled by specializing
the algorithm and not by generating the constraints
and adding these to the main primal constraints
.
Copyright © 2008 by SAS Institute Inc., Cary, NC, USA. All rights reserved.