Page 111 - Innovations in Intelligent Machines
P. 111

UAV Path Planning Using Evolutionary Algorithms  101
                           selection of centers, which is the most generalized form of a RBFN since the
                           location of the centers undergo a supervised learning process along with the
                           rest of the network’s free parameters.
                              The standard process is to select the input vectors in the training set as
                           RBFN centers. In this case results M=NR, where NR is the number of train-
                           ing data. For large training sets (resulting in large M values) this choice is
                           expected to increase storage requirements and CPU cost. Additionally, the
                           M=NR choice could lead to over-fitting and/or bad generalization of the
                           network. The proposed solution [49, 45] is the selection of M<NR and con-
                           sequently the search for sub-optimal solutions, which will provide a better
                           generalizing capability to the network.
                              As far as training is concerned, there are two different approaches, the
                           direct and the iterative learning. In our case the first approach was adopted.
                           The direct learning process is based on a matrix formulation of the governing
                           equations of RBF network. The presentation of the network with the NR input
                           patterns allows the formulation of a (NR × M) matrix H, which becomes
                           square in the special case when NR=M. Each line in the interpolation matrix
                           H corresponds to a learning example and each column to a RBFN center. The
                           output unit values result in the form of the matrix product:

                                           H (NR × M) w (M × 1) = yy (NR × 1) ,            (29)
                           where yy is the desired output vector as provided by the training dada set,
                           and w is the synaptic weights vector, which consists of M unknowns to be
                           computed.
                              A possible way for inverting H is through the Gram-Schmidt technique.
                           H is first decomposed as
                                                        H = QR,                            (30)
                           with Q and R being (NR × M) and (M × M) matrices respectively, where R
                           is upper triangular and
                                                   T
                                                  Q Q = diag (1, 1,... , 1) .              (31)
                           After the computation of Q and R matrices, the weights vector can be com-
                           puted using back-substitution in

                                                             T
                                       R(M × M)w(M × 1) = Q (M × NR)yy(NR × 1)             (32)
                              There are several reasons why one should choose RBFN as the approxi-
                           mation model; Haykin [49] offers comparative remarks for RBFNs and Multi-
                           layer Perceptrons (MLPs). However, the main reason for choosing RBFNs is
                           their compatibility with the adopted local approximation strategy, as it is
                           described in the subsequent section. The use of relatively small numbers of
                           training patterns i.e. small networks, helps creating local range RBFNs. That
                           in turn allows the inversion of matrix H to use almost negligible CPU time
                           and the approximation error is kept very small. We should keep in mind that
   106   107   108   109   110   111   112   113   114   115   116