Page 69 - Rapid Learning in Robotics
P. 69

4.4 Learning Phases                                                                      55


                 Assuming that a proper subset of input components I has been specified
                 (usually when the number of input components is equal to the mapping
                 dimension, jIj   m), an input vector Px       can be completed with a zero


                 distance (E s  =0) best-match location Pw     s     Px in the
                                                                               mapping man-
                 ifold M. Then, only the output components are effectively adapted by
                 Eq. 4.14 (equivalent to the selection matrix      P). If the measurement of
                 the training vector components is differently reliable or noisy, the learning
                 parameter   might be replaced by a suitable diagonal matrix controlling
                 the adaptation strength per component (we return to this issue in Sect. 8.2.)

                     Fig. 4.8 illustrates this procedure for an m     dimensional three-node
                 PSOM. We consider the case of a d            dimensional embedding space

                  x           x and I   f g, i.e., a mapping x     x     x . The embedded mani-

                 fold M is drawn before and after 300 adaptation steps (using Eq. 4.14 with
                 a learning steps parameter   linearly decreased from 0.8 to 0.1) together
                 with the reference vectors and the noisy training data set.

                     Note, that Eq. 4.14 is taking the entire PSOM output w s   into account,
                 in contrast to w a in the “locally acting” Kohonen step of Eq. 3.10. But the
                 alternatively proposed learning rule (Ritter 1993)


                                          	w a         a  s    x H  w a                  (4.15)

                 turns out to be unstable. A correctly trained PSOM is even corrupted by a
                 sequence of further training examples.
                     We demonstrate this fact by an example. For the same learning task,
                 Fig. 4.9 depicts the initially correct parabola shaped mapping manifold
                 M and the asterisk marked locations of the three reference vectors w a in
                 the front curve. The other curves, shifted to the back, show M after sev-
                 eral epochs of 25 adaptation steps, each using Eq. 4.15 (uniformly random
                 sampled training vectors,              ). The adaptation leads to a    displace-
                 ment of the reference vectors in such a way, that the symmetric parabola
                 becomes distorted. This shows that Eq. 4.15 is not able to maintain the
                 correct mapping. The reference vectors are shaken around too-much – in a
                 non-localized manner, and in both the output and input space. Note, that
                 the mapping manifold M in Fig. 4.9 does not display one simple parabola
                 but in fact the parameterized curve  w    s        s w (the joint ordinate pro-

                 jection of two parabolas w    s        s  with
                                                             w the invisible abscissa s in the
                 parameter manifold S; Initially w    s  is constructed here as a linear func-
                 tion (degenerated parabola) in s, similar as Fig. 4.8).
   64   65   66   67   68   69   70   71   72   73   74