Page 323 - Classification Parameter Estimation & State Estimation An Engg Approach Using MATLAB
P. 323

312                                      WORKED OUT EXAMPLES

            The five-fold cross-validation errors of ldc and qdc are 13.0% and
            17.4%, respectively. Note that by using the function crossval,we
            avoided having to use the training set for testing. If we had done that,
            the errors would be 11.9% and 16.2%, but this estimate would be
            biased and too optimistic. The results also depend on how the data is
            randomly split into batches. When this experiment is repeated, you
            will probably find slightly different numbers. Therefore, it is advisable
            to repeat the entire experiment a number of times (say, 5 or 10) to get
            an idea of the variation in the cross-validation results. However, for
            many experiments (such as the feature selection and neural network
            training below), this may lead to unacceptable training times; there-
            fore, the code given does not contain any repetitions. For all the
            results given, the standard deviation is about 0.005, indicating that
            the difference between ldc and qdc is indeed significant. Notice that
            we use the word ‘significant’ here in a slightly loose sense. From a
            statistical perspective it would mean that for the comparison of the
            two methods, we should state a null-hypothesis that both methods
            perform the same, and define a statistical test (for instance a t-test) to
            decide if this hypothesis holds or not. In this discussion we use the
            simple approach in which we just look at the standard deviation of
            the classifier performances, and call the performance difference sig-
            nificant when the averages differ by more than two times their stand-
            ard deviations.
              Even with these simple models on the raw, not preprocessed data, a
            relative good test error of 13.0% can be achieved (with respect to the
            simplest approach by looking at the class probabilities). Note that qdc,
            which is a more powerful classifier, gives the worst performance. This is
            due to the relatively low sample size: two covariance matrices, with
            1  13(13 þ 1) ¼ 91 free parameters each, have to be estimated on 80%
            2
            of the data available for the classes (i.e. 172 and 233 samples, respect-
            ively). Clearly, this leads to poor estimates.



            9.1.3  Feature extraction

            It might be expected that more flexible methods, like the  -nearest neigh-
            bour classifier (knnc, with   optimized using the leave-one-out method)
            or the Parzen density estimator (parzenc) give better results. Surpris-
            ingly, a quick check shows that then the errors become 19.6% and 21.9%
            (with a standard deviation of about 1.1% and 0.6%, respectively), clearly
   318   319   320   321   322   323   324   325   326   327   328