Page 230 - Artificial Intelligence in the Age of Neural Networks and Brain Computing
P. 230
2. The Neural Network Approach 221
come from other neurons and its output is distributed to other nodes. The organiza-
tion of the neurons is hierarchical if the nodes are structured in layers. This kind of
topology is somehow reminiscent of the organization of pyramidal neurons in the
mammalian brain. Other kinds of topologies have been considered, in particular,
the maps and the grids, where the neurons are organized in 2D or 3D distributions.
The most typical neural network architecture is referred to as multilayer perceptron
(MLP): here the neurons are organized in successive layers. A pictorial illustration
of the MLP NN is reported in Fig. 11.1A. The MLP is quite similar to the ADALINE
previously proposed in signal processing literature apart from the presence of the
nonlinearities at least in the hidden layer’s nodes. Various nonlinear functions
have been proposed for approximation, pattern recognition, and classification
problems. In MLP the nodes in successive layers are connected and the connections
are weighted. Among them monotonic saturating functions (i.e., sigmoids) are a fa-
vorite choice. Thus, the MLP is of feedforward type. However, in recursive networks
there are also feedback links and this enriches the dynamical behavior of NNs [1].
The interest of MLP in approximating inputeoutput mappings has been moti-
vated by the proofs of some theorems demonstrating that any continuous mapping
can be approximated by MLP with at least one hidden layer whose output functions
are sigmoid (or monotonic) functions [2]. It is worth mentioning that this notion of
universal approximation just states that the NN can learn, not what in practice it
really does learn.
NNs are adaptive systems that are trained aiming to derive an optimal represen-
tation of the weights’ matrices. The training is carried out through a specific
“learning” procedure. The learning can be supervised (SL), unsupervised (UL),
semisupervised (SSL), or reinforced (RL). In SL, NNs are forced to associate their
outputs to real (or complex) valued targets fixed by a “teacher,” through a procedure
(typically gradient-based) that optimizes approximation errors (a “cost” function).
FIGURE 11.1
(A) A shallow (one hidden layer) and (B) a deep (multiple hidden layers) neural network.