Page 248 - Compact Numerical Methods For Computers
P. 248

Conjugate gradients method in linear algebra        235
                      In some tests run by S G Nash and myself, the inexact line searches led to very
                      slow convergence in a number of cases, even though early progress may have
                      been rapid (Nash and Nash 1977).

                                19.2. SOLUTION OF LINEAR EQUATIONS AND
                          LEAST-SQUARES PROBLEMS BY CONJUGATE GRADIENTS
                      The conjugate gradients algorithm 22 can be modified to solve systems of linear
                      equations and linear least-squares problems. Indeed, the conjugate gradients
                      methods have been derived by considering their application to the quadratic form
                                                        T
                                            S(b)=½b Hb-c b+(anyscalar)                 (15.11)
                                                   T
                      where H is symmetric and positive definite. The minimum at b' has the zero
                      gradient
                                                   g =Hb' - c =0                       (15.17)

                      so that b' solves the linear equations
                                                      Hb' = c.                           (19.1)
                      The linear least-squares problem
                                                                                         (19.2)
                      can be solved by noting that A A is non-negative definite. For the present
                                                     T
                      purpose we shall assume that it is positive definite. Hestenes (1975) shows that it
                      is only necessary for H to be non-negative definite to produce a least-squares
                      solution of (19.1). Thus identification of
                                                      H=A A                            (19.34
                                                           T
                                                           T
                                                       c=A f                           (19.3b)
                      permits least-squares problems to be solved via the normal equations.
                        The particular changes which can be made to the general-purpose minimising
                      algorithm of §§16.1 and 16.2 to suit these linear systems of equations better are
                      that (a) the linear search and (b) the gradient calculation can be performed
                      explicitly and simply. The latter is accomplished as follows. First, if the initial point
                      is taken to be                    ( 0 )
                                                       b   =0
                                                                                        (19.4)
                      then the initial gradient. from (15.17), is
                                                      g, = -c.                          (19.5)

                      If the linear search along the search direction t  yields a step-length factor k , then
                                                                                        j
                                                              j
                      from (15.18)
                                                   g  =g +k Ht .                         (19.6)
                                                    j +l  j  j  j
                      This with (19.5) defines a recurrence scheme which avoids the necessity of
                      multiplication of b by H to compute the gradient, though Ht j  must be formed.
                      However, this is needed in any case in the linear search to which attention will
                      now be directed.
   243   244   245   246   247   248   249   250   251   252   253