Page 57 - Introduction to Statistical Pattern Recognition
P. 57

2 Random Vectors and their Properties                         39




                                        hI#O,  hZ= ...= hn=O,                  (2.141)

                                       -
                              ih;=h - tr(ATMMTA) = tr(M*AA*M)  = M7X-'M  ,     (2.142)
                                     I
                              i=l
                     where AA * = C-' is obtained from (2. IOl), and tr(MTZ-'M) = MrC-'M  because
                     MTZ-l M is a scalar. Thus,

                                          IS1 = lZl(1 +MYM).                   (2.143)


                         Small sample size problem: When only m samples are available in an n-
                                                                             A
                     dimensional vector space with m < n, the sample autocorrelation matrix S is calcu-
                     lated from the samples as

                                                                               (2.144)


                     That is,  is a function of m or less linearly independent vectors. Therefore, the
                     rank of ,?  should be m or less. The same conclusion can be obtained for a sample
                                                    ,.
                     covariance matrix. However, the (Xi - M)'s are not linearly independent, because
                                             A
                     they are related by E? (Xi - M) = 0.  Therefore, the rank of a sample covariance
                                      I  =I
                     matrix is (m - I) or less. This problem, which is called a small sample size prob-
                     lem, is often encountered in pattern recognition, particularly when n is very large.
                     For this type of problem, instead of calculating eigenvalues and eigenvectors from
                     an n x n matrix, the following procedure is more efficient [9].
                          Let XI, . . . , X,  (m c n) be samples. The sample autocorrelation matrix of
                     these samples is

                                                                               (2.145)


                     where U,,,   is called a sample matrix and defined by
                                            u = [XI . *  Xnilnxm .             (2.146)
                                                    .
                     Instead of using the n x n matrix i,lx, of (2.143,  let us calculate the eigenvalues
                     and eigenvectors of an m x m matrix (U TU)mml as
                                        1
                                       -(~~~)mmtQmmi = Q'nixniAmxni  .         (2.147)
                                        m
                     Multiplying U into (2.147) from the left side, we obtain
   52   53   54   55   56   57   58   59   60   61   62