Page 24 - Biosystems Engineering
P. 24

Micr oarray Data Analysis Using Machine Learning Methods       5


               FIGURE 1.2  Details  x 1
               of a neuron.             w 11

                                      w 12      net 1
                                 x 2                 f(.)  o 1

                                          w 1n


                                  x n

               shown in the figure, the inputs to the network are x , x ,…, x , and the
                                                          1  2   n
               network output is y. Figure 1.2 illustrates the details of a single neuron
               where net , known as the activation level, is the sum of the weighted
                       1
               inputs to the neuron and f(.) represents an activation function, which
               transforms the activation level of a neuron into an output signal. Typi-
               cally, an activation function could be a threshold function, a sigmoid
               function (an S-shaped, symmetric function that is continuous and dif-
               ferentiable), a gaussian function, or a linear function. For example, the
               output of the neuron in Fig. 1.2 (i.e., the output of the first neuron in the
               first hidden layer of the network in Fig. 1.1) can be written as follows:
                                              ⎛  n   ⎞
                                  o =  f net =(  )  f ⎜∑ w x j⎟
                                   1      1        j 1  ⎠
                                              ⎝ =
                                               j 1
               where the synaptic weights w , w ,…, w  define the strength of con-
                                        11  12   1n
               nection between the neuron and its inputs. Such synaptic weights
               exist between all pairs of neurons in each successive layer of the net-
               work. They are adapted during learning to yield the desired outputs
               at the output layer of the network.
                   A multilayer feedforward network, whose neurons in the hidden
               layers have sigmoidal activation functions, is known as a multilayer
               perceptron (MLP) network. MLP networks are capable of learning
               complex input–output mapping. That is, given a set of inputs and
               desired outputs, an adequately chosen MLP network can emulate the
               mechanism that produces data through learning.
                   In a  supervised learning paradigm, the network uses training
               examples (specifically, desired outputs for a given set of inputs) to
               determine how well it has learned and to guide adjustments of the
               synaptic weights in order to reduce its overall error. An example of a
               supervised learning rule is the back-propagation algorithm (Rumel-
               hart and McClelland 1986), which is developed to train MLP networks
               based on the principle of steepest gradient method. The training of a
               neural network is complete when a prespecified stopping criterion is
               fulfilled. A typical stopping criterion is the performance of the net-
               work on a validation dataset, which is a portion of the training sam-
               ples that was not used for updating the weights.
   19   20   21   22   23   24   25   26   27   28   29