Page 25 - Biosystems Engineering
P. 25

6    Cha pte r  O n e

                   In contrast to supervised learning, unsupervised learning discovers
               patterns or features in the input data with no help from a teacher,
               essentially performing a clustering of the input space. Typical unsu-
               pervised learning rules include Hebbian learning, principal compo-
               nent learning, and Kohonen learning rules. These rules are applied in
               training self-organizing neural network architectures such as self-
               organizing maps (SOM). Although neural networks developed for
               both the supervised and unsupervised learning paradigms have per-
               formed very well in their respective application fields, improvements
               have been developed by combining the two paradigms. Examples
               include a radial basis function (RBF) network and learning vector
               quantization (LVQ) discussed as follows.
                   RBF networks are special feedforward networks that have a single
               hidden layer. The activation functions of the neurons in the hidden
               layer are radial basis functions, whereas the neurons in the output
               layer have simple linear activation functions. Radial basis functions
               are a set of predominantly nonlinear functions such as gaussian
               functions that are built up into one function. Each gaussian function
               responds only to a small region of the input space where the gauss-
               ian is centered. Thus, while an MLP network uses hyperplanes
               defined by weighted sums as arguments to sigmoidal functions, an
               RBF network uses hyperellipsoids to partition the input space in
               which measurements or observations are made. Thus, RBF networks
               find the input-to-output map using local approximators. The key to
               a successful implementation of RBF networks is to find suitable cen-
               ters for the gaussian functions. Theoretically, the centers and width
               of the gaussian functions can be determined with supervised learn-
               ing, for example, the steepest gradient method. They can also be
               determined through unsupervised learning by clustering the training
               data points. Once the centers and width of the gaussian functions are
               obtained, the weights for the connection between the hidden layer
               and the output layer are easily and quickly determined using meth-
               ods such as linear least squares, as the output neurons are simple
               linear combiners.
                   LVQ selects training vectors with a known classification and pres-
               ents them to the network to examine cases of misclassification. An
               LVQ network has a first competitive layer and a second linear layer.
               The competitive layer consists of reference vectors that learn to cate-
               gorize input vectors. The linear layer transforms the competitive
               layer’s classes into target classifications defined by the users. During
               training, a class labeled “input vector” (e.g., expression pattern) is
               picked at random and is compared with each reference vector. The
               input vector is assigned to a class that the most similar reference vec-
               tor possesses. The reference vector is moved closer to the input vector
               if the assignment was correct; otherwise it is moved away. During
               operation, the unknown expression profiles are classified based on
               their similarity to the reference vectors.
   20   21   22   23   24   25   26   27   28   29   30