Page 172 -
P. 172

160      5 Neural Networks


                               want  w'xi > 0 for  patterns  xi belonging  to  class  W, and  w7xi < 0 for  patterns xi
                               belonging  to class  @. These two conditions can be  written simply as w7xifi > 0,
                               using  the  target  values.  The  fact  that  wlxiti must  be  positive  for  correct
                               classification,  suggests  the  use  of  the  following  error  function  known  as  the
                               perceptron criterion:

                                  E(w) =-  Xw7xiti .
                                         xi in error

                                 Let us look more closely at this error function. Consider the situation with d = 2
                               depicted in Figure 5.11, where the unit length normal to the discriminant, n, points
                               in the positive direction (d(x) > 0), corresponding to class w,.
                                 Let  us  determine  the  projection  distance  1 - lo of  a feature vector  x  onto the
                               positive  normal  n.  The projection  of  d(x) referred  to  the  origin,  lo, is  given by
                               -wdllnll  '.  The length I  is the projection of x  onto n. Since the components of  the
                               normal vector n to the linear discriminant d(x) are precisely the weights of d(x), n
                               = (w,, w2)  j,  we can compute the projection distance l(x) of  x  onto the normal of
                               the linear discriminant. as:







                                 This  projection  I(x) is positive for feature vectors x  lying in  the positive half
                               plane and negative otherwise.
                                  Consider  now  that  feature  vector  x  represented  in  Figure  5.1 1  is  wrongly
                               classified  because  it  has  target  value  -1.  As  the  target  value  is  negative,  the
                               contribution of the pattern to the error (5-14) is a positive value. In the same way, a
                               feature  vector  lying  in  the  negative  half  plane,  therefore  with  I(x)  < 0, will
                               contribute  positively  to  the  error  if  its  target  value  is  +1.  In  general,  the
                               contributions  of  the  wrongly  classified  patterns  to  the  errors  are  the  Euclidian
                               distances to the discriminant.
                                  The perceptron compensates for these errors by applying the following learning
                               rule:

                               - Pattern correctly classified: do nothing.
                               - Pattern wrongly classified: add the pattern to the weight vector if the pattern is
                                  from class CU~ (ti = + 1) and subtract it if it is from class w  (ti = - 1).

                                  Hence, the increment of the weight vector is:

                                  Aw=tixi  for  wrong  xi.                                    (5-16)



                                 See expressions (2-2d) relative  to  the  distance of  d(x) from  the  origin  and  the  unitary
                                 normal vector pointing into the positive direction.
   167   168   169   170   171   172   173   174   175   176   177