Page 187 - Introduction to Statistical Pattern Recognition
P. 187

4  Parametric Classifiers                                    169



                    ferent from the ones used for test (given Xi's), the algorithm of  (3.1 19)-(3.128)
                    must be used to calculate the theoretical error.


                    4.4  Other Classifiers

                         In this section, we will discuss subjects which were left out in the previ-
                    ous discussions.  They are the piecewise classifiers and some of  the properties
                    in binary inputs.


                    Piecewise Classifiers

                         If  we  limit  our  discussion  to  two-class  problems,  quadratic  or  linear
                    classifiers have wide applications.  However, when we  have to handle three or
                    more classes, a  single quadratic or  linear classifier cannot be  adopted effec-
                    tively.  Even  in  two-class problems, the  same is  true when  each class consists
                    of several clusters.  For these cases, a set of classifiers, which is called a piece-
                    wise classifier, gives increased flexibility.
                         Piecewise quadratic for multiclass problems: For multiclass problems,
                    the multihypothesis test in the Bayes sense gives the best classifier with regard
                    to minimizing the error.  That is, from (3.44)

                                    Pkpp(X) = max Pipi(X) +  X  E  q .         (4.147)
                                              I
                    If the distributions of X for L classes are normal, (4.147) is replaced by

                                  1                      1
                              min[-(X  - M,)'z;'(x  - M,) + - ln I  I - w,]  ,   (4.148)
                               i 2                      2

                    where max  is  changed to  min  because of  the  minus-log operation.  Note  that
                    the  normalized distance of X  from each class mean, Mi, must be  adjusted by
                    two constant terms, (112)ln  ICi I and In Pi. Equation (4.148) forms a piecewise
                    quadratic, boundary.

                         Piecewise quadratic for multicluster problems: For multicluster prob-
                    lems, the boundary is somewhat more complex.  Assuming that L  = 2, and that
                    each distribution consists of  m, normal clusters with  the cluster probability of
                    Pi, for the jth cluster, the Bayes classifier becomes
   182   183   184   185   186   187   188   189   190   191   192