Page 203 - Artificial Intelligence in the Age of Neural Networks and Brain Computing
P. 203

2. Emergence of Some Neural Network Principles     193




                  strength if stimuli were unpaired [85]. In fact, the findings on long-term potentiation
                  at synapses were shortly followed by other findings on long-term depression
                  [16e18].
                     A synthesis of Hebb and McCulloch-Pitts occurred in various versions of the
                  perceptrons due to Rosenblatt [19] whose multilayer perceptrons were trained to
                  respond selectively to specific patterns of sensory stimuli. Perceptrons formed the
                  basis for the widely used back propagation networks [4,20], which in turn form
                  the basis for the currently popular deep learning methods [21e23].
                     The mathematics employed in the McCulloch-Pitts and Rosenblatt models was
                  discrete, as befits a network of all-or-none neurons. Yet the biophysics of neurons,
                  like other physical processes, points to the importance of variables that can take
                  on a continuum of possible values. Rashevsky [24] and other pioneers showed
                  that the discrete processes in a network of neurons could be averaged over a large
                  number of neurons to generate continuous processes in networks whose units are
                  populations rather than individual neurons.



                  2. EMERGENCE OF SOME NEURAL NETWORK PRINCIPLES
                  From about the mid-1960s to the mid-1980s, when much less-detailed cognitive
                  neuroscience results were available than are now, several pioneering neural
                  modelers developed network architectures based largely on abstract principles,
                  through incorporating some of that period’s state of the art in neuroscience
                  (cf. Ref. [1], Chapters 3 and 4). This led to the development of some “building
                  blocks” that could be incorporated into larger and more complex networks. Since
                  different building blocks were used to model different functions, this was not a
                  “one size fits all” approach to neural modeling.
                     One of these principles, inspired both by Hebb and by psychologists such as Hull
                  [13], was associative learning. Associative learning was a prominent feature in the
                  work of Grossberg, Anderson, Kohonen, and Kosko [25e38]. Grossberg’s early
                  networks, such as the outstar, were particularly notable for including conditioned
                  decreases as well as increases in synaptic strength, in order that a network could learn
                  the proportions between activations of different units in a stimulus pattern.
                     Just as associative learning is important in models of long-term memory, lateral
                  inhibition or competition plays a crucial role in models of short-term memory. This
                  principle is particularly important in models of perception [39e43].Competitive
                  interactions capture data showing that the visual system does not store every stimulus
                  it receives equally but sharpens contrast and reduces noise.
                     Opponent processing is a psychological principle that captures the fact that people
                  and animals often perceive contrasts rather than absolute amounts of variables.
                  Contrast effects were first observed in the realm of emotion: absence of electric shock
                  when one has just been shocked is rewarding, and absence of food when one is
                  hungry and expecting food is punishing. Grossberg ([5]), explained these effects
                  using the gated dipole model with neural transmitter accumulation and depletion,
   198   199   200   201   202   203   204   205   206   207   208