The Interior Point Nonlinear Programming Solver -- Experimental |
Overview of Constrained Optimization
A function that plays a pivotal role in establishing conditions that characterize a local minimum of an
NLP problem is the Lagrangian function
, defined as

Note that the Lagrangian function can be seen as a linear combination of the objective and constraint
functions. The coefficients of the constraints,

, and

, are called the Lagrange
multipliers or dual variables. At a feasible point

, an inequality constraint is called active if it
is satisfied as an equality, i.e.,

. The set of active constraints at a feasible point

is then defined as the union of the index set of the equality constraints,

, and the indices of those inequality constraints
that are active at

, i.e.,

An important condition that is assumed to hold in the majority of optimization algorithms is the so-called
linear independence constraint qualification (LICQ). The LICQ states that at any feasible point

, the
gradients of all the active constraints are linearly independent. The main purpose of the LICQ is to ensure that
the set of constraints is well-defined, in a way that there are no redundant constraints, or in a way that there are
no constraints defined such that their gradients are always equal to zero.
The First-Order Necessary Optimality Conditions
If

is a local minimum of the NLP problem and the LICQ holds at

, then there are vectors of Lagrange
multipliers

and

, with components

, and

, respectively,
such that the following conditions are satisfied:

where

is the gradient of the Lagrangian function with respect to

, defined as

The preceding conditions are often called the
Karush-Kuhn-Tucker (KKT) conditions , named after their inventors.
The last group of equations (

) is called the complementarity condition. Its main
aim is to try to force the Lagrange multipliers,

, of the inactive inequalities
(i.e., those inequalities with

) to zero.
The KKT conditions describe the way the first derivatives of the objective and constraints are related at a local
minimum
. However, they are not enough to fully characterize a local minimum. The second-order optimality
conditions attempt to fulfill this aim, by examining the curvature of the Hessian matrix of the Lagrangian function
at a point that satisfies the KKT conditions.
The Second-Order Necessary Optimality Condition
Let

be a local minimum of the NLP problem, and let

be the corresponding Lagrange multipliers that
satisfy the first-order optimality conditions. Then

for all
nonzero vectors

that satisfy the following conditions:
-
, 
-
,
, such that 
-
,
, such that 
The second-order necessary optimality condition states that, at a local minimum, the curvature of
the Lagrangian function along the directions that satisfy the preceding conditions must be nonnegative.