Page 964 - The Mechatronics Handbook
P. 964

0066_Frame_C32.fm  Page 14  Wednesday, January 9, 2002  7:54 PM









                       For a neuron in the input layer to be reacting just for the stored pattern, the threshold value for this
                       neuron should be

                                                           ) =  n 1)
                                                       w n+1(   ( –  –                          (32.35)
                       If it is required that the neuron must also react for similar patterns, then the threshold should be set to
                       w n +1  = −[n − (1 + HD)], where HD is the Hamming distance defining the range of similarity. Since for
                       a given input pattern only one neuron in the first layer may have the value of 1 and remaining neurons
                       have 0 values, the weights in the output layer are equal to the required output pattern.
                         The network, with unipolar activation functions in the first layer, works as a lookup table. When the
                       linear activation function (or no activation function at all) is used in the second layer, then the network
                       also can be considered as an analog memory. For the address applied to the input as a binary vector, the
                       stored set of analog values, as weights in the second layer, can be accurately recovered. The feedforward
                       counterpropagation network may also use analog inputs, but in this case all input data should be
                       normalized,

                                                        w i =  x ˆ i =  --------                (32.36)
                                                                  x i
                                                                  x i
                         The counterpropagation network is very easy to design. The number of neurons in the hidden layer
                       is equal to the number of patterns (clusters). The weights in the input layer are equal to the input patterns,
                       and the weights in the output layer are equal to the output patterns. This simple network can be used
                       for rapid prototyping. The counterpropagation network usually has more hidden neurons than required.
                       However, such an excessive number of hidden neurons are also used in more sophisticated feedforward
                       networks such as the probabilistic neural network (PNN) Specht (1990) or the general regression neural
                       networks (GRNN) Specht (1992).

                       WTA Architecture

                       The winner take all (WTA) network was proposed by Kohonen (1988). This is basically a one-layer
                       network used in the unsupervised training algorithm to extract a statistical property of the input data,
                       Fig. 32.14(a). At the first step, all input data are normalized so that the length of each input vector is the
                       same and, usually, equal to unity, Eq. (32.36). The activation functions of neurons are unipolar and
                       continuous. The learning process starts with a weight initialization to small random values. During the
                       learning process the weights are changed only for the neuron with the highest value on the output—the
                       winner:

                                                       ∆w w =  c xw w )                         (32.37)
                                                              (
                                                                 –
                       where
                         w w = weights of the winning neuron,
                          x = input vector,
                          c = learning constant.

                       Usually, this single-layer network is arranged into a two-dimensional layer shape, as shown in Fig. 32.14(b).
                       The hexagonal shape is usually chosen to secure strong interaction between neurons. Also, the algorithm
                       is modified in such a way that not only the winning neuron but also neighboring neurons are allowed
                       for the weight change. At the same time, the learning constant c in Eq. (32.37) decreases with the distance
                       from the winning neuron. After such an unsupervised training procedure, the Kohonen layer is able to
                       organize data into clusters. Output of the Kohonen layer is then connected to the one- or two-layer
                       feedforward network with the error backpropagation algorithm. This initial data organization in the
                       WTA layer usually leads to rapid training of the following layer or layers.


                       ©2002 CRC Press LLC
   959   960   961   962   963   964   965   966   967   968   969