Page 57 - Neural Network Modeling and Identification of Dynamical Systems
P. 57

2.1 ARTIFICIAL NEURAL NETWORK STRUCTURES                  45





















                          FIGURE 2.13 Examples of a structural organization for feedforward dynamic neural networks. (A) NARX (Nonlinear
                          AutoRegression with eXternal inputs). (B) LDDN (Layered Digital Dynamic Network). D in aresource(input) data; D out are
                          output data (results); L (0)  is input layer; L (1)  is hidden layer; L (2)  is output layer for NARX network and hidden layer for
                          LDDN; L (3)  is output layer for LDDN; TDL (m) , TDL (m) , TDL (n1) , TDL (n2)  are tapped delay lines of order m, m, n1,and n2
                                                         1     2      1      1
                          respectively.










                                                                       FIGURE 2.15 General structure of the operational (hid-
                                                                                              (0)
                                                                       den and output) ANN layers: s  is the ith neuron of the
                                                                                              i
                                                                       pth ANN layer; W(L (p) ) is a matrix of synaptic weights for
                                                                       connection entering the neurons of the L (p)  layer.

                                                                                   0
                                                                       inputs; and a ∈ R is the value of the ith input.
                                                                                   i
                                                                       For each ith neuron of an lth layer we denote
                                                                                     l
                                                                       the following: n is the weighted sum of neu-
                                                                                     i
                                                                                  l
                                                                       ron inputs; ϕ : R → R is the activation function;
                                                                                  i
                                                                            l
                                                                       and a ∈ R is the output of an activation function
                                                                            i
                          FIGURE 2.14 ANN input layer as a data structure.  (the neuron state). Outputs a of activation func-
                                                                                                L
                                                                  (0)                           i
                          (A) One-dimensional array. (B) Two-dimensional array. s  ,
                                                                  i    tions of Lth layer neurons are called the network
                          (0)
                          s  are numeric or character variables.
                          ij                                           outputs. Also, W ∈ R n w  is the total vector of net-
                                                                                                             l
                                                                       work parameters, which consists of biases b ∈ R
                                                                                                             i
                          Neural Network (LFNN). We introduce the fol-  and connection weights w l i,j  ∈ R. Thus, the lay-
                          lowing notation: L ∈ N is the total number of  ered feedforward neural network is a parametric
                                 l
                          layers; S ∈ N is the number of neurons within  function family, mapping the network inputs a 0
                                                                                                       L
                                       0
                          the lth layer; S ∈ N is the number of network  and parameters W to the outputs a according
   52   53   54   55   56   57   58   59   60   61   62