Page 201 - Neural Network Modeling and Identification of Dynamical Systems
P. 201

192           5. SEMIEMPIRICAL NEURAL NETWORK MODELS OF CONTROLLED DYNAMICAL SYSTEMS

                         Algorithm 1 Simple homotopy continuation            5.5 OPTIMAL DESIGN OF
                         training algorithm for semiempirical ANN                 EXPERIMENTS FOR
                         model (5.2).                                       SEMIEMPIRICAL ANN-BASED
                                    ¯
                         Require: δ, δ,  τ  min ,  τ                                     MODELS
                           1: a ∼ U(W)
                           2: w ← a
                                                                         The indirect approach to acquisition of train-
                           3: τ ← 0                                   ing data sets for ANN-based models of dynam-
                           4: while τ< 1 and  τ >  τ min  do
                                                                      ical systems, described in Section 2.4.3, can also
                           5:   ˜ τ ← min{τ +  τ,1}
                                                                      benefit from theoretical knowledge of the sim-
                           6:   ˜ w ← LM(E,a,w,τ)
                                                                      ulated system. Recall that we need to design a
                                            ¯
                           7:   if   ˜ w − w  < δ then
                                                                      set of reference maneuvers that maximize the
                           8:      w ← ˜ w
                                                                      resulting training set representativeness. Such a
                                       τ
                           9:      τ ←˜
                                                                      set of maneuvers might be designed manually
                          10:      if   ˜ w − w  <δ then
                                                                      by an expert in the specific domain, although
                          11:          τ ← 2 τ                        this procedure is quite time consuming and the
                          12:      end if                             results tend to be suboptimal. Methods for au-
                          13:   else                                  tomation of this procedure constitute the sub-
                                         1
                          14:       τ ←  τ
                                         2                            ject of study for optimal design of experiments
                          15:   end if                                [41]. Classical theory of optimal design of exper-
                          16: end while
                                                                      iments is mostly dedicated to linear regression
                                                                      models. Extensions of this theory to active selec-
                                                                      tion of most informative training examples for
                         function (5.45) with respect to parameters w,  feedforward neural networks were suggested
                         while keeping τ fixed. It uses the current pa-  in [42,43]. More recently, these results were ex-
                         rameter values as initial guess. The Levenberg–  tended to active selection of controls that pro-
                         Marquardt method is denoted by LM in the algo-  vide most informative training examples for re-
                         rithm description. The continuation algorithm  current neural networks [44]; however, the pri-
                         also involves some form of step length adapta-  mary focus is on the greedy optimization with
                                                                      one-step-ahead prediction horizon. All of the
                         tion, whereby if the norm of model parameters
                         change exceeds δ, the predictor step length  τ is  abovementioned methods alternate between the
                                        ¯
                         decreased and the corrector step is reevaluated.  following three steps: search for the most infor-
                                                                      mative training examples to include in the data
                         Conversely, if the norm of model parameters  set guided by the current model estimate; acqui-
                         change does not exceed δ, the step length is in-
                                                                      sition of the selected training examples; retrain-
                         creased. The initial guess for parameter values a
                                                                      ing or adaptation of the model using the new
                         is picked at random.
                                                                      training set. Since this approach relies on the
                            Note that a conceptually similar approach of
                                                                      specific form of the model and involves model
                         solving a series of problems with increasing pre-  training after inclusion of each training exam-
                         diction horizon was suggested in [37–40], and  ple, it is better suited for online adaptation of the
                         it has proved to be highly successful. Results of  existing model rather than the design of a new
                         computational experiments with this algorithm  model from scratch.
                         for training of semiempirical ANN-based mod-    In this section we discuss an approach to
                         els of a maneuverable F-16 aircraft motion are  the optimal design of reference maneuvers for
                         presented in Chapter 6.                      semiempirical neural network–based models of
   196   197   198   199   200   201   202   203   204   205   206