Page 156 - Computational Retinal Image Analysis
P. 156

150    CHAPTER 8  Image quality assessment




                         quality for a subset of the data of the main study. The Messidor data set was not
                         included as the number of ungradable images were low. Analysis was reported for
                         the EyePACS-1 dataset. Eight ophthalmologists were involved in grading the 9963
                         EyePACS-1 images using a grading tool that recorded overall image quality. The
                         image quality was assessed according to criteria relating to focus, illumination, image
                         field definition and artifacts. For fully gradable images the sensitivity and specificity
                         were reported as 93.9% and 90.9%, respectively. AUC was reported as 0.978.
                            Additionally, the algorithm performance in detecting diabetic retinopathy was also
                         assessed in terms of its performance in handling mydriatic and non-mydriatic images
                         in the EyePACS-1 dataset, which may affect image quality. The authors report similar
                         performance for mydriatic images compared to non-mydriatic images and these
                         findings could impact screening protocols with respect to routine use of mydriasis.

                         2.4.2   Human visual system information combined with convolutional
                         neural networks
                         A further implementation of deep learning was reported by Mahapatra et al. [40] that
                         was aimed specifically at providing retinal image quality assessment and taking into
                         account the role of the human visual system (HVS) by the inclusion of saliency map
                         information. Saliency maps aim to define where a particular region is different (or
                         noticeable) from its neighbors with respect to image features. Itti et al. [55] previously
                         developed a visual attention system to select conspicuous locations of a scene. The
                         saliency model proposed combines multiscale feature maps to form a local saliency
                         map which takes both local and global features into account that are relevant for
                         image quality assessment and the identification of gradable and ungradable images.
                            In addition to information from the saliency maps, information from a trained
                         convolutional neural network (CNN) was included in the final image quality
                         assessment system. The CNN included five convolution layers and five max pooling
                         layers. Rectified Linear Units were used to facilitate faster training and to reduce
                         sensitivity to the scale of the input. During training negative log-likelihood was used
                         as the loss function and stochastic gradient descent using dropout was implemented
                         to reduce overfitting. Data augmentation was applied to expand the training set by
                         image translation and horizontal reflections to expand the dataset size by 50 times.
                         The training data comprised 80% of the total dataset size.
                            To produce the image quality classification, feature vectors derived from the saliency
                         maps were used to train one random forest (RF) classifier. A second feature vector
                         derived from the last fully connected layer of the CNN was used to train a second RF
                         classifier. Both classifiers used the same image labels. A combination of the probability
                         outputs from the two RF classifiers was used to give a final classification of “gradable”
                         or “ungradable”, therefore combining both supervised and unsupervised image features.
                            The algorithm was evaluated against three other techniques and the data set was
                         acquired from a DR screening program and included 9653 ungradable images and
                         11,347 gradable images. All images were non-mydriatic and were graded by human
                         graders. The sensitivity was reported as 98.2%, with a specificity of 97.8%. The
                         system was reported to have low computation time and therefore could contribute
   151   152   153   154   155   156   157   158   159   160   161