Professional Documents
Culture Documents
267
Consequently, in this chapter we will discuss five major approaches for solving nonlinear programming problems with constraints:
1. Analytic solution by solving the first-order necessary conditions for optimality
(Section 8.2)
2. Penalty and barrier methods (Section 8.4)
3. Successive linear programming (Section 8.5)
4. Successive quadratic programming (Section 8.6)
5. Generalized reduced gradient (Section 8.7)
The first of these methods is usually only suitable for small problems with a few
variables, but it can generate much useful information and insight when it is applicable. The others are numerical approaches, which must be implemented on a computer.
+ X2
+ X;
- 1=0
Solution. This problem is illustrated graphically in Figure E8. la. Its feasible region is
a circle of radius one. Contours of the linear objective x, + x2 are lines parallel to the
one in the figure. The contour of lowest value that contacts the circle touches it at the
point x* = (-0.707, -0.707), which is the global minimum. You can solve this problem analytically as an unconstrained problem by substituting for x, or x2 by using the
constraint.
Certain relations involving the gradients off and h hold at x* if x* is a local rninimum. These gradients are
Vf(x*)
[1?1]
where A*
= 0.
PART
FIGURE ES.la
Circular feasible region with objective function contours
and the constraint.
FIGURE ES.lb
Gradients at the optimal point and at a nonoptimal point.
The relationship in Equation (a)must hold at any local optimum of any equalityconstrained NLP involving smooth functions. To see why, consider the nonoptimal
point x1in Figure E8.lb. Vflxl)is not orthogonal to the tangent plane of the constraint
at xl, so it has a nonzero projection on the plane. The negative of this projected gradient is also nonzero, indicating that moving downward along the circle reduces
269
(improves) the objective function. At a local optimum, no small or incremental movement along the constraint (the circle in this problem) away from the optimum can
improve the value of the objective function, so the projected gradient must be zero.
This can only happen when Vf(x*) is orthogonal to the tangent plane.
Vf ( x * ) + A* Vh(x*) = 0
where A* = 0.707. We now introduce a new function L(x, A) called the Lagrangian
function:
L(x, A ) = f ( x )
+ Ah(x)
(8.6)
vxL(x9A)lb*.r*)= 0
so the gradient of the Lagrangian function with respect to x, evaluated at (x*, A*),
is zero. Equation (8.7),plus the feasibility condition
constitute the first-order necessary conditions for optimality. The scalar A is called
a Lagrange multiplier.
x:+x;-1=0
The first-order necessary conditions for this problem, Equations (8.9)-(8.1 l ) , consist of three equations in three unknowns (x,, x,, A). Solving (8.9)-(8.10) for x, and
x2 gives
1
270
PART
which shows that x, and x2 are equal at the extremum. Substituting Equation (8.12)
into Equation (8.1 1);
and
XI
= x2 = ? 0.707
The minus sign corresponds to the minimum off, and the plus sign to the maximum.
(a)
Subject to:
+ 5xi
h ( x ) = 0 = 2x, + 3x2 - 6
(b)
(c)
Solution. Let
L(x,A)
4x:
Source: Edgar, Himmelblau and Lasdon. Optimization of chemical processes, 2nd ed.