Professional Documents
Culture Documents
AX = B,
1
y= 2 XAX BX
y = 0 = AX B.
Hence the vector X0 that minimizes Equation (3.2.21) will also be the solution
vector of equation (3.2.20). We shall proceed to develop the method based
upon the minimization of the quadratic form given by Equation (3.2.21). The
extension of the algorithm to the optimization of a more general objective
function will then be discussed.
The idea behind the conjgate-gradient procedure is similar to that of
steepest descent in that a sequence of one-dimensional searches is carried out
in directions which are determined by the partial derivaties of the objective
function. Unlike the method of steepest, however, the search vectors are not
equal to the negative gradient vectors; rather, a sequence of search vectors is
determined in such a manner that each search vector. The algorithm is
guaranteed to minimize a quadratic function of n independent variables with no
more tan n iterations a conditionknown as quadratic convergence.
Proceeding from an arbitrary initial search point X 0, we locate a sequence
of points that are successively closer to the mnimum as follows:
(3.2.23)
Xi+1 = Xi + iPi,
where i is a positive scalar defines the distance between X i and Xi+1 along the
search vector Pi. Notice that t the minimum along Pi will occur where Pi is
yi+1Pi= Pi yi+1 = 0.
Xk = Xi
In particular,
Subtracting X0 from each side of of Equation (3.2.26) and premultiplyng by A,
we obtain
Let us now develop a criterion for defining the search vector P j. Premultiplyng
Equation (3.2.29) by Pi-1 gives
The first termo n the right-hand side vaniches because of Equation (3.2.24). If
we now choose the Pj such that
PAPj = 0
For i j, then the summation term in Equation (3.2.31) will also vanish, so that
i unless yn = 0. We see, then, that the condition that the vectors P, be Aconjugate causes yn to vanish identically. Since this condition exists only
where the quadratic form is minimized, we conclude that the quadratic form is
thus minimized after no more than n one-dimensional searches in the
directions P0, P1, , Pn-1.
We still have not specified exactly how the Pi are chosen. Le tus
arbitrarily let P0 = -y0, and the let
(3.2.34)
Where rhe i are positive scalars that must be determined, This choice of
Pi can be shown to satisfy the A-conjugacy condition expressedby Equation
(3.2.32). For further decelopment of this point the reader is referred to
Beckman [1960].
From Equation (3.2.32), we can write
(3.2.35)
PiAPi+1 = 0
Again referring to Equation (3.2.24), we see that last term in Equation (3.2.40)
vanishes. Hence,
So that
yi+1yi = 0.
Xi+1 = Xi + iPi
As the point on the P, vector where the objective function is extremized. This
point is located by conducting a one-dimensional search along P i. Determine
yi+1, the gradient vector, at Xi+1. Compute i in accordance with Equation
(3.2.52), and determine a new search vector
(3.2.53)
Again the plus sign is chosen for a maximization problema and the minus sign
corresponds to a minimization.
Although the method is supposed to find the optimum of a quadratic
form with no more than n iterations, it is in fact rather sensitive to roundoff
error. Fletcher and Reeves [1964] suggest that the computation be restarted
with Pi =yi after every n+1 iterations as an effective way to minimize the
problema of cumulative roundoff.
Our discussion thus far has been concerned exclusively with the
optimization of quadratic forms. We can see that the method is applicable to a
broader clase of function. However by expanding the objective function in a
Taylor series:
Notice that H is a real , symmetric matrix, providing the ogjective fuction is not
linear.
EXAMPLE 3.2.4
Reslove Example 3.2.1 using the method of conjugate gradients, with X 10 = 1
and X20 = 1 as an initial point.
In vector notation,
X0 =
[]
1
1
[ ]
4
72
So that
[]
1 +
1
X1 =
[ ]
4
72
> 0.
) = (4
- 2)2 + 9(72
), we obtain y = 3.1594 at
X1 =
1.223 .
5.011
3.554
0.197
as follows:
- 4)2.
=0.0555. Hence
[ ]
y|x1 =
0 =
= 0.00244.
P1 =
Solving for
[ ][
] yields y= 5.91x10-10 at
X2 =
and
= 0.4986. Hence
3.0000
,
5.000
X1n
0.5000
0.1800
0.1703
0.1922
0.1977
0.1998
0.19998
X2n
0.5000
0.7716
0.7618
0.7264
0.6984
0.6685
0.66683
yn
5.7899
7.1772 x 10-3
3.8202 x 10-3
1.2420 x 10-3
2.9656 x 10-4
4.0192 x 10-7
5.0831 x 10-12
Where
Where
PiAPj = 0,
ij.
(3.2.63)
Any algorithm that satisfies Equation (3.2.63) will, apart from numerical
roundoff errors, minimize a quadratic form with no more than n onedimensional searches in the Pi directions, i = 0,1,,n-1.
Thus far the description of the variable metric algorithm parallels that of
the methos of conjugate gradients. The algorithms differ in the manner in
which the search vectors P1, P2,, Pn-1 are chosen.
In the variable metric algorithm the search vectors are chosen as