Page 85 - Introduction to Statistical Pattern Recognition
P. 85

3  Hypothesis Testing                                         67



                     I-@)  = 41(x) + . . . + qn.-~(X) + qk+l(X) + . . .  + qL(X) = 1 - qk(X), and  the
                     Bayes error is the expected value of r(X) over X.  That is,

                                   r(X) = 1 - max qi(X)  and   E = E(r-(X)} .   (3.45)
                                             I
                          When cost functions are involved, the decision rule becomes

                                        rk(X) = min ri(X)  +  X  E  ok          (3.46)
                                                I
                     where ri(X) is a simple extension of (3.20) to L classes as
                                                    L
                                             ri(X> = Ccijqj(X)                  (3.47)
                                                    j=l
                     and  ci,  is  the cost of  deciding X  E  wi when X  E  mi.  Substituting (3.47) into
                     (3.46) and using the Bayes theorem,
                                  L               L
                                 CckjPjp,(X) = min  xcijPjpj(X)  +  X  E  ok .   (3.48)
                                 j=l           i   j=1
                     The resulting conditional cost given X  and the total cost are

                                    r(X) = min rj(X)  and   I‘  = E(r(X)} .     (3.49)
                                            I

                          Example 5:  When cii = 0 and cij = 1 for i#j, ri(X) of (3.47) becomes

                                                                                (3.50)


                     Therefore, the  decision  rule  of  (3.46)  and  the  resulting  conditional cost  of
                     (3.49) become (3.43) and (3.43, respectively.

                     Single Hypothesis Tests

                          So far, we have assumed that our task  is to classify an unknown sample
                     to one of L classes.  However, in practice, we often face the problem in  which
                     one class is well defined while the others are not.  For example, when we want
                     to  distinguish targets  from  all  other possible  nontargets, the  nontargets may
                     include trucks, automobiles, and all kinds of other vehicles as well as trees and
                     clutter discretes.  Because of  the wide variety, it is almost impossible to study
                     the distributions of all possible nontargets before a decision rule is designed.
   80   81   82   83   84   85   86   87   88   89   90