Page 173 - Handbook of Deep Learning in Biomedical Engineering Techniques and Applications
P. 173

162   Chapter 6 Plant leaf disease classification based on feature selection




                                       24@48x48
                                                       24@16x16
           8@128x128                                                 1x256
                            8@64x64
                                                                                    1x128
                                                                                            1x4






                            Max-Pool      Convolution      Max-Pool      Dense
                                        Figure 6.6 An example of a CNN.


                                    the other. The preprocessing required in a ConvNet is much
                                    lower as compared with other classification algorithms. While
                                    in primitive methods, filters are hand-engineered, with enough
                                    training, ConvNets can learn these filters or characteristics
                                    (Fig. 6.6).
                                       In this study, we explore three popular CNN architectures,
                                    namely, AlexNet, VGG16, and ResNet to build the classification
                                    system.
                                    3.3.1 AlexNet (2012)
                                       AlexNet [14] was an innovative architecture to tackle large
                                    labeled data sets for image recognition with higher precision
                                    and efficiency. AlexNet introduced various key design innovations
                                    like the addition of dropout layers for higher accuracy along with
                                    incorporation of distributed processing for better scalability and
                                    faster training. Hence, leveraging a multi-GPU system can speed
                                    up training as well as evaluation for large data sets. AlexNet sets
                                    the premise to better image classification architectures and
                                    research using DL techniques and methods. Fig. 6.7 shows the
                                    architecture of AlexNet.
                                       The architecture for the AlexNet contains five convolution
                                    neural layers and three fully connected (FC) layers. The architec-
                                    ture of AlexNet used ReLU activation function in its neural layers
                                    to accommodate faster training over traditional activation func-
                                    tions like tanh. The response to normalization of layer’s average
                                    data given during training to prevent stagnant learning
                                    iterations and high false positives in recognition. The max pool-
                                    ing layers helped reduce variance also capturing strong inputs
                                    over the network layers. Pooling layers are placed after the
                                    response normalization layers. The architecture uses overlapping
                                    pooling in its structure.
   168   169   170   171   172   173   174   175   176   177   178