Page 185 - Artificial Intelligence in the Age of Neural Networks and Brain Computing
P. 185

3. From RNNs to Mouse-Level Computational Intelligence: Next Big Things  175




                  the kind of recurrence one would use to represent models like simultaneous equation
                  models in economics, or implicit models defined by some kind of relaxation to equi-
                  librium. Feedforward CoNNs can approximate a wide range of smooth functions, but
                  for effective intelligent control and true brain-like intelligence it is important to have
                  a more general capability to approximate nonsmooth functions. Our 2008 paper gave
                  an example of a simple learning problem which arises in learning how to navigate a
                  cluttered space, as a general skill (not just learning a particular maze). The usual
                  CoNN simply could not learn such a difficult task, but the CSRN could. The
                  paper also described a further extension beyond CSRN suitable for exploiting
                  non-Euclidean symmetries, which is essential for true brain-like capabilities.
                     For an even more powerful universal learning system able to capture all the ben-
                  efits of the most advanced TLRN design and extensions of CSRN, one would simply
                  combine both types of recurrence in a single system (as does the brain). See the
                  Handbook of Intelligent Control [17] for discussion of how to combine both types
                  together.


                  3.2 DEEP VERSUS BROAD: A FEW PRACTICAL ISSUES
                  Philip Chen, a recent president of the IEEE Society for Systems, Man and Cyber-
                  netics, has argued this year [28] that “broad” neural networks could perform just
                  as well as “deep” (many layered) neural networks. In actuality, debates about the
                  number of layers were already resolved in principle by 1990, when I explained
                  [14] how to use the more general static ANN design illustrated in Fig. 8.11:
                     Formally, this is the deepest possible feedforward network possible with N
                  neurons and implementation on a sequential computer. Each neuron is a layer
                  unto itself, making it all broad as well as deep. All possible feedforward network
                  structures with N neurons are a special case of this general structure. The key is
                  to use rational, statistically grounded learning (using penalty functions, pruning
                  and growing methods) to learn the exact details. Such general methods have been


















                  FIGURE 8.11
                  Generalized MLP design.
   180   181   182   183   184   185   186   187   188   189   190