Page 68 - Classification Parameter Estimation & State Estimation An Engg Approach Using MATLAB
P. 68

BAYESIAN ESTIMATION                                           57

              . On the other hand, if the prior knowledge is weak, i.e. if the
                variances of the parameters are very large, the inverse covariance
                matrix C  1  tends to zero. In the limit, the estimate becomes:
                        x

                                           T   1     1  T   1
                             ^ x x MMSE ðzÞ¼ H C H  H C z              ð3:21Þ
                                             v          v
                In this solution, the prior knowledge, i.e. m , is completely ruled out.
                                                      x
              Note that the mode of a normal distribution coincides with the expec-
            tation. Therefore, in the linear-Gaussian case, MAP estimation and
            MMSE estimation coincide: ^ x MMSE (z) ¼ ^ x MAP (z).
                                                x
                                      x

            3.1.4  Maximum likelihood estimation

            In many practical situations the prior knowledge needed in MAP estima-
            tion is not available. In these cases, an estimator which does not depend
            on prior knowledge is desirable. One attempt in that direction is the
            method referred to as maximum likelihood estimation (ML estimation).
            The method is based on the observation that in MAP estimation, (3.17),
            the peak of the first factor p(zjx) is often in an area of x in which the
            second factor p(x) is almost constant. This holds true especially if
            little prior knowledge is available. In these cases, the prior density p(x)
            does not affect the position of the maximum very much. Discarding
            the factor, and maximizing the function p(zjx) solely, gives the ML
            estimate:

                                ^ x x ML ðzÞ¼ argmaxfpðzjxÞg           ð3:22Þ
                                            x
            Regarded as a function of x the conditional probability density is called the
            likelihood function. Hence the name ‘maximum likelihood estimation’.
              Another motivation for the ML estimator is when we change our
            viewpoint with respect to the nature of the parameter vector x. In the
            Bayesian approach x is a random vector, statistically defined by means
            of probability densities. In contrast, we may also regard x as a non-
            random vector whose value is simply unknown. This is the so-called
            Fisher approach. In this view, there are no probability densities asso-
            ciated with x. The only density in which x appears is p(zjx), but here x
            must be regarded as a parameter of the density of z. From all estimators
            discussed so far, the only estimator that can handle this deterministic
            point of view on x is the ML estimator.
   63   64   65   66   67   68   69   70   71   72   73