Page 319 - Introduction to Statistical Pattern Recognition
P. 319
7 Nonparametric Classification and Error Estimation 301
A similar trend was observed in the parametric cases of Chapter 5, but the
trend is more severe with a nonparametric approach. These problems are
addressed extensively in this chapter.
Both Parzen and kNN approaches will be discussed. These two
approaches offer similar algorithms for classification and error estimation, and
give similar results. Also, the voting kNN procedure is included in this
chapter, because the procedure is very popular, although this approach is
slightly different from the kNN density estimation approach.
7.1 General Discussion
Parzen Approach
Classifier: As we discussed in Chapter 3, the likelihood ratio classfier
is given by -InpI(X)/p2(X) ><r, where the threshold t is determined in various
ways depending on the type of classifier to be designed (e.g. Bayes, Neyman-
Pearson, minimax, etc.). In this chapter, the true density functions are replaced
by their estimates discussed in Chapter 6. When the Parzen density estimate
is
with a kernel function IC,(.) used, the likelihood ratio classifier becomes
where S = {X\’), . . . ,X$!,X\2), . . . ,X$! } is the given data set. Equation (7.1)
classifies a test sample X into either o1 or 02, depending on whether the left-
hand side is smaller or larger than a threshold t.
Error estimation: In order to estimate the error of this classifier from
the given data set, S, we may use the resubstitution (R) and leave-one-out (L)
methods to obtain the lower and upper bounds for the Bayes error. In the R
method, all available samples are used to design the classifier, and the same
sample set is tested. Therefore, when a sample Xi” from o1 is tested, the fol-
lowing equation is used.