Page 390 - Numerical Methods for Chemical Engineering
P. 390

Linear least-squares regression                                     379



                                  T
                  parameter values. X X is a real, symmetric matrix that is at least positive-semidefinite. For
                  a well-designed set of experiments that provides sufficient data to estimate each parameter
                                             T
                  to at least some finite accuracy, X X is positive-definite.


                  Solving the least-squares linear system
                         T
                  Often, X X may be ill conditioned, i.e., it has one or more eigenvalues near zero, when the
                  experimental design does not provide sufficient information to measure well one or more
                  linear combinations of parameters. Therefore, QR decomposition is the favored solution
                  method,

                                                     T
                                                    X X = QR
                                                                                     (8.25)
                                         T
                                               −1
                                        Q = Q        R is upper-triangular
                                                      T
                                            T
                  As Q is orthogonal, we write (X X)θ LS = X y as
                                                  T             T  T
                                        QRθ LS = X y ⇒ Rθ LS = Q X y                 (8.26)
                  The solution of the latter system requires only backward substitution.
                    A more expensive approach, but one that provides further insight into experimental
                                                             T
                  design, is to diagonalize the real, symmetric matrix X X:
                                                 T
                                                X X = V  V  T                        (8.27)
                                         T
                  V is an orthogonal matrix, V = V −1 , whose columns are the normalized eigenvectors of
                                                                                       T
                    T
                  X X and   is a diagonal matrix whose principal diagonal contains the eigenvalues of X X.
                  The least-squares estimate θ LS is
                                                             T
                                                          T
                                                      −1
                                             θ LS = [V   V ]X y                      (8.28)
                                    T
                                              T
                  Note also that since X Xθ LS = X y, we have an overdetermined system,
                                                  Xθ LS = y                          (8.29)
                  that can be solved by SVD (Chapter 3). The SVD approach is popular; however, here we
                  use the familiar eigenvalue decomposition (8.27) in lieu of SVD to analyze experimental
                  designs.


                  Example. Least-squares fitting of rate law parameters to transformed
                  batch data

                  We now apply the least-squares method to the transformed batch reactor data of Table 8.2
                  for the kinetics of A + B → C, using the model

                                                                                     (8.30)
                                   log r R1 = log k 1 + ν a log c A + ν b log c B
                                      10       10         10         10
   385   386   387   388   389   390   391   392   393   394   395