Page 330 - Computational Statistics Handbook with MATLAB
P. 330
Chapter 9: Statistical Pattern Recognition 319
it belongs to. We build a classifier using these data and (possibly) one of the
techniques that are described in this chapter. To use the classifier, we measure
the four features for an iris of unknown species and use the classifier to assign
the species membership.
Sometimes we are in a situation where we do not know the class member-
ship for our observations. Perhaps we are unable or unwilling to assume how
many groups are represented by the data. In this case, we are in the unsuper-
vised learning mode. To illustrate this, say we have data that comprise mea-
surements of a type of insect called Chaetocnema [Lindsey, Herzberg, and
Watts, 1987; Hand, et al., 1994]. These variables measure the width of the first
joint of the first tarsus, the width of the first joint of the second tarsus, and the
maximal width of the aedegus. All measurements are in microns. We suspect
that there are three species represented by these data. To explore this hypoth-
esis further, we could use one of the unsupervised learning or clustering tech-
niques that will be covered in Section 9.5.
9.2 Bayes Decision Theory
The Bayes approach to pattern classification is a fundamental technique, and
we recommend it as the starting point for most pattern recognition applica-
tions. If this method is not adequate, then more complicated techniques may
be used (e.g., neural networks, classification trees). Bayes decision theory
poses the classification problem in terms of probabilities; therefore, all of the
probabilities must be known or estimated from the data. We will see that this
is an excellent application of the probability density estimation methods from
Chapter 8.
We have already seen an application of Bayes decision theory in Chapter 2.
There we wanted to know the probability that a piston ring came from a par-
ticular manufacturer given that it failed. It makes sense to make the decision
that the part came from the manufacturer that has the highest posterior prob-
ability. To put this in the pattern recognition context, we could think of the
part failing as the feature. The resulting classification would be the manufac-
) that sold us the part. In the following, we will see that
turer (M A or M B
Bayes decision theory is an application of Bayes’ Theorem, where we will
classify observations using the posterior probabilities.
We start off by fixing some notation. Let the class membership be repre-
,
,
, j = 1 … J for a total of J classes. For example, with the iris
sented by ω j
data, we have J = 3 classes:
ω 1 = Iris setosa
ω 2 = Iris versicolor
ω 3 = Iris virginica.
© 2002 by Chapman & Hall/CRC

