Page 234 - Artificial Intelligence in the Age of Neural Networks and Brain Computing
P. 234

3. Deep Architectures and Learning    225
























                  FIGURE 11.3
                  Stacked autoencoders architecture. The first autoencoder (AE 1 ) maps the input instance
                  x into a compressed representation h 1 (coding operation) which is used to reconstruct the
                  input data (decoding operation). After training AE 1 , the code h 1 is used as input to train
                  AE 2 , providing the code vector h 2 and so on. The procedure is repeated for all AE n
                  autoencoders. The compressed representations h 1 , h 2 . h n form the stacked architecture
                  (SAE) which is typically fine-tuned using conventional back propagation algorithm.

                  the visual cortex of the brain and have been widely applied in image and speech
                  recognition. A CNN includes an automatic feature extractor (composed by multiple
                  stages of convolution and pooling layers) and a standard MLP-NN which processes
                  the features learned before for classification tasks (Fig. 11.4) [7]. The convolutional

















                  FIGURE 11.4
                  CNN architecture. It includes a sequence of convolution (CONV) and pooling (POOL) layers
                  followed by a standard fully connected neural network. In the convolutional layer the input
                  map convolves with K filters (or kernels), providing K feature maps. After applying a
                  nonlinear activation function (sigmoidal or a rectified linear unit) to each feature map, the
                  pooling layer is performed. The features learned are the input of a fully connected neural
                  network followed by a softmax layer, which performs the classification tasks.
   229   230   231   232   233   234   235   236   237   238   239