Page 367 - Introduction to Statistical Pattern Recognition
P. 367

7  Nonparametric Classification and Error Estimation          349



                         One comment is  in  order regarding the  application of  Option 4  to  kNN
                    estimation.  In  Step 2  of  Option 4  of  the  Parzen  case,  it  is  fairly  simple  to
                    remove the effect of Xf) (the test sample) from the density estimates of  all the
                    other samples using (7.58).  There  is no analogous simple modification  in  the
                    kNN  case.  In  order to  remove  the  effect  of  Xf) from  all  other  density  esti-
                    mates, one must remove Xf’  from the table of  nearest neighbors, rearrange the
                    NN  table, and  recalculate all  of  the  density estimates.  This  procedure would
                    have to  be repeated to test each of  the samples in the design set, resulting in a
                    fairly drastic increase in computation time.  In practice, modifying each of  the
                    density  estimates to  remove  the  effect  of  Xf) is  not  nearly  as  important  as
                    finding the threshold by  minimizing the error among the remaining N, + Nz-1
                    samples.  That is, modifying the estimates of  the likelihood ratios in Step 2 is
                    not  necessary to  get  reliable results  -  we  do  it  in  the  Parzen  case  primarily
                    because  it  is  easy.  Thus  for  kNN  estimation,  Step  2  of  Option  4  involves
                    finding  and  sorting L(X:))  for  all  samples  Xy) # Xf), finding  the  value  of  t
                    which  minimizes  the  error  among  these  Nl+N2-1  samples,  and  using  this
                    value oft to classify xj,?.
                         Figure 7-14 shows the  results of  applying Option 4  to the kNN  estima-
                    tion problem.  For comparison, the results obtained using t = 0 are also shown.




                         Experiment 14: Estimation of the kNN error, L and R
                               Same as Experiment 4, except
                               Metric: A I = C I  and A  = C2  (Instead of kernel)
                               No. of  neighbors: k  = 1-30  (Instead of kernel size)
                               Threshold: Option 4 and t  = 0
                               Results: Fig. 7-14 [12]


                    As  in  the  Parzen  case, the  threshold plays  its  most  significant role  when  the
                    covariances  of  the  data  are  different,  and  particularly  when  the  covariance
                    determinants are different.  In  Data I-I, the bias of  the density estimates for o,
                    and  w2 are  nearly  equal  near  the  Bayes  decision  boundary,  and  hence  good
                    results  are  obtained  without  adjusting  the  threshold.  However,  for  Data  I-41
                    and I-A, the kNN  errors are heavily biased  and unusable without adjusting the
                    threshold.
   362   363   364   365   366   367   368   369   370   371   372