Page 302 - Introduction to Statistical Pattern Recognition
P. 302
284 Introduction to Statistical Pattern Recognition
Experiment 3: A similar experiment was conducted for a double
exponential waveform as
(6.1 18)
where three parameters are uniformly distributed in
0.7 5 a 5 1.3 ,
0.3 5 m 5 0.7 , (6.1 19)
0.3 5 z 10.6 .
Using eight sampling points and 250 waveforms, the intrinsic dimensionality
of the data was estimated, and the results are shown in Table 6-4. Again,
fairly accurate estimates of the intrinsic dimensionality (which is 3) were
obtained.
Experiment 4: The intrinsic dimensionalities of Data RADAR were
estimated by (6.115). They were found to be 19.8 for Chevrolet Camaro and
17.7 for Dodge Van, down from the original dimensionality of 66. This indi-
cates that the number of features could be reduced significantly. Although this
technique does not suggest how to reduce the number of features, the above
numbers could serve as a guide to know how small the number of features
should be.
Very Large Number of Classes
Another application in which the kNN distance is useful is a
classification scenario where the number of classes is very large, perhaps in the
hundreds. For simplicity, let us assume that we have N classes whose expected
vectors Mi (i = l,,.,,N) are distributed uniformly with a covariance matrix I,
and each class is distributed normally with the covariance matrix 0~1.
When only a pair of classes, 0; and ai, considered, the Bayes
is
classifier becomes a bisector between Mi and M,i, and the resulting error is
m
Ep = j 1 e-.'ZIZ dx (pairwise error) , (6.120)
d(M,.M,)iZo
where d(Mj,Mj) is the Euclidean distance between Mi and Mj. Equation
(6.120) indicates that E,, depends only on the signal-to-noise ratio, d(Mi,Mj)l~.
When the number of classes is increased, Mi is surrounded by many neighbor-
ing classes as seen in Fig. 6-5, where MmN is the center of the kth nearest