Page 242 - Artificial Intelligence in the Age of Neural Networks and Brain Computing
P. 242

5. Deep Learning Models for EEG Signal Processing   233




                   (A)                              (B)
                                        Deep Learning
                                                                          Deep Learning
                        Feature                          Feature   Data    Other
                                      Data-Driven
                       Engineering
                                                        Engineering  Driven  Features
                                 Features
                                       Features
                         Step                             Step    Features  Extractor

                                      Classifica on
                                                                 Classifica on
                                         Step
                                                                   Step
                  FIGURE 11.11
                  (A) Serial and (B) parallel DL schemes.


                  features form a unique input vector for the classification step. The serial step is
                  preferred for reducing the information redundancy of the features. In a recent
                  paper [30], a serial scheme based on SAEs has been proposed that includes a
                  time-frequency transformation of the input recordings, an intermediate step of
                  data-driven feature combination, and a final classification stage. The SAE model
                  has been proposed for discriminating EEGs of subjects affected by early stage
                  CJD from other forms of rapidly progressive dementia (RPD).
                     Each AE is implemented by an MLP-NN that includes an encoding stage fol-
                  lowed by a decoder. Higher-order features have been obtained by stacking two levels
                  of nonlinear (sigmoidal) nodes. The outputs of the deepest hidden layer is the final
                  feature vector used as input of the classification step. The processing chain is
                  depicted in Fig. 11.12. It is worth noting that while the “engineered” features are
                  extracted channel by channel, the higher-order features generated by the SAE are
                  mixing information pertaining to all of the channels. This procedure can be limited
                  to selected areas, that is, frontal or parietal channels, in order to gain information on
                  brain areas mostly relevant for the classification.
                     The output of each hidden layer of the SAE is given by:
                                              h ¼ sðF xÞ;

                  where s is the node nonlinearity, for example, the standard sigmoidal function:
                                                      z  1

                                            s ¼ 1 þ e     .
                  F is the learned matrix of the encoding layer of the MLP-NN, and z ¼ F x. The loss/
                  cost function to be minimized through learning is given by:


                                         L ¼ 
e x   x 
 þ l F  2

                                                   2
   237   238   239   240   241   242   243   244   245   246   247