Page 157 - Computational Colour Science Using MATLAB
P. 157

144                CHARACTERIZATION OF PRINTERS






























               Figure 9.1 Schematic diagram for an MLP. The solid lines represent weighted connections
               between the processing units (*)





               the current set of weights) is computed, and the values of the weights are
               modified to reduce this error. This process is repeated for each input–output pair
               in the training set and the presentation of the whole training set in this way is
               known as a training epoch. Training may require thousands or even hundreds of
               thousands of epochs and typically the training procedure is very computationally
               intensive. However, at the end of the training period the values of the weights are
               fixed. During the testing mode, input vectors are presented to the network and
               output vectors are computed. The performance of the network in testing mode
               using the data from the training set is known as the training error. A common
               problem with MLPs is that they are prone to over-fitting the training data. As the
               number of hidden layers or units in the network increases, the training error
               should decrease. In the limit a sufficiently complex MLP can produce a training
               error of zero; such a network, however, may exhibit poor generalization
               performance. Generalization is the ability of the network to perform using data
               that was not used during the training period. A second data set, known as a
               testing data set, is therefore used to determine the testing error. Of course, the
               training and testing data sets should be drawn from the same population so that
               they both represent, in a statistical sense, the problem being addressed by the
               network.
   152   153   154   155   156   157   158   159   160   161   162