Page 251 -
P. 251

Exerciscs   239


                             Rosenblatt F (1962) Principles of Neurodynamics.  Spartan Books. Washington DC.
                             Ruinelhart DE, Hinton GE, Williams RJ  (1986) Learning Internal  Representations by  Error
                               Propagation.  In:  Rumelhart  DE,  McClelland  JL  (eds)  Parallel  Distributed  Processing:
                               Explorations in the Microstructure of Cognition, vol.1, chapter 8, MIT Press.
                             Simon  HU  (1997)  Bounds on  the  Number  of  Examples  Needed  for Learning  Functions.
                               SIAM J. of Computing, 26:751-763.
                             Specht DF (1990) Probabilistic  Neural Networks. Neural Networks,  13: 109-1 18.
                             Specht DF (1991) A Generalized Regression  Neural  Network.  IEEE Tr Neural  Networks,
                               21568-576.
                             Vapnik VN (1998) Statistical Learning Theory. Wiley, New York.
                             Van  Rooij  AJF,  Jain  LC,  Johnson  RP  (1996)  Neural  Network  Training  Using  Genetic
                               Algorithms. World Scientific Co. Pte. Ltd., Singapore.
                             Vonk E, Jain LC, Johnson RP (1997) Automatic Generation of Neural Network Architecture
                               Using Evolutionary  Computation. World Scientific Co. Pte. Ltd., Singapore.
                             Weigend AS, Rumelhart DE, Huberman  BA (1991) Generalization by  Weight-Elimination
                               with  Application  to  Forecasting.  In:  Lippman  RP,  Moody  JE,  Touretzky  DS  (eds)
                               Advances  in  Neural  Information  Processing  Systems,  32375-882, Morgan  Kaufmann,
                               California.
                             Widrow B, Glover Jr JR, McCool JM, Kaunitz J, Williams CS, Hearn RH, Zeidler JR, Dong
                               Jr E.  Goodlin RC  (1975) Adaptive Noise Cancelling:  Principles and Applications Proc
                               IEEE, 63: 1692-1716.
                             Widrow B, Hoff Jr M (1960) Adaptive  Switching Circuits. In: IRE WESCON Conv. Rec.,
                               4:96-104.



                             Exercises

                             5.1  Consider the adaptive noise cancelling application  of a linear network, as in  thc ECG
                                example described in section  5.1. Let R = E[xx'] represent the auto-correlation matrix
                                of the signal fed at the network inputs.
                                a)  Compute the error expectation E[q], noticing  that  the error for the sample  input
                                    vector xi is q= $ - w'xj. Show that the Hessian matrix of the error energy is equal
                                    to R.
                                 b)  Taking into account formula 5-42, prove that an upper bound of the learning factor
                                    I],  for the 50 Hz noise cancelling example of section 5.1 is q,,,, = 4/(n2~, where a
                                    is  the  amplitude  of  the  sinusoidal  inputs.  Check  this  upper  bound  using  the
                                    ECG5OHz.xls file.
                                 c)  The time constant of an exponential approximating the mean-square error learning
                                    curve is given by  r=w1(41] tr(R)), where tr(R) is the sum of the eigenvalues of R.
                                    Show that for the pure sinusoidal noise cancelling example in section 5.1, the time
                                    constant is r=1/(2na2q).

                              5.2  Determine  the  optimal  parabolic  discriminants  for  the  two-class  one-dimensional
                                 problems  shown  in  Figure  5.8  and  Figure  5.9  by  solving  the  respective  normal
                                 equations.

                              5.3  Implement  the  first  two  steps  of  the  gradient  descent  approach  for  the  parabolic
                                 discriminant of  the two-class  one-dimensional  problems  shown in  Figure 5.9,  starting
   246   247   248   249   250   251   252   253   254   255   256