Page 802 - Mechanical Engineers' Handbook (Volume 2)
P. 802

2 Background   793

                                 Dendrites

                                                                           Axon terminals
                                           Nucleus
                                                             Node of Ranvier
                                                     Myelin




                                                        Axon             Synapses
                                           Cell body


                           Figure 1 Nervous system cell. (With permission from http://www.sirinet.net/ jgjohnso/index.html.)


            2   BACKGROUND
            2.1  Neural Networks

                           The multilayer NN is modeled based on the structure of biological nervous systems (see
                                                                              n
                                                                                                 m
                           Fig. 1) and provides a nonlinear mapping from an input space R into an output space R .
                           Its properties include function approximation, learning, generalization, classification, and so
                           on. It is known that the two-layer NN has sufficient generality for closed-loop control pur-
                           poses. The two-layer NN shown in Fig. 2 consists of two layers of weights and thresholds
                           and has a hidden layer and an output layer. The input function x(t) has n components, the
                           hidden layer has L neurons, and the output layer has m neurons.
                              One may describe the NN mathematically as
                                                              T
                                                                   T
                                                         y   W  (Vx)
                           where V is a matrix of first-layer weights and W is a matrix of second-layer weights. The
                           second-layer thresholds are included as the first column of the matrix W by augmenting
                                                                                      T
                           the vector activation function  ( ) by 1 in the first position. Similarly, the first-layer thresh-
                                                                     T
                           olds are included as the first column of the matrix V by augmenting vector x by 1 in the
                           first position.
                              The main property of NNs we are concerned with for control and estimation purposes
                           is the function approximation property. 2,3  Let ƒ(x) be a smooth function from R → R . Then,
                                                                                             m
                                                                                        n
                           it can be shown that if the activation functions are suitably selected and is restricted to a
                           compact set S   R , then for some sufficiently large number L of hidden-layer neurons, there
                                         n
                           exist weights and thresholds such that one has
                                                                 T
                                                             T
                                                     ƒ(x)   W  (Vx)    (x)
                           with  (x) suitably small. Here,  (x) is called the neural network functional approximation
                           error. In fact, for any choice of a positive number   , one can find a NN of large enough
                                                                     N
                           size L such that  (x)    for all x   S.
                                               N
                              Finding a suitable NN for approximation involves adjusting the parameters V and W to
                           obtain a good fit to ƒ(x). Note that tuning of the weights includes tuning of the thresholds
                           as well. The neural net is nonlinear in the parameters V, which makes adjustment of these
                           parameters difficult and was initially one of the major hurdles to be overcome in closed-
   797   798   799   800   801   802   803   804   805   806   807