Page 148 - Computational Retinal Image Analysis
P. 148

142    CHAPTER 8  Image quality assessment




                         et al. [16] applied a convolutional neural network to automatically grade diabetic
                         retinopathy and diabetic macular oedema in retinal fundus images. The algorithm
                         was also assessed in terms of its performance in judging image quality.
                            All the algorithms described in this section aim to provide an indication of image
                         quality in relation to various applications. Images are either divided into distinct
                         classes (e.g. excellent, good etc.) or quantify image quality (usually on a scale from
                         0 to 1). The processing speeds of the algorithms described above is variable. Methods
                         that use generic image quality often have low computational requirements, while
                         methods that use structural parameters have higher requirements. Image quality
                         assessment related to epidemiological studies usually take place off-line once the
                         entire dataset has been captured. In applications where image quality assessments are
                         required at the point of capture, such as in diabetic retinopathy screening systems,
                         real-time operation could enhance the accuracy of automated analysis algorithms
                         such as the automated recognition of microaneurysms.


                         2.2  Datasets and metrics used to evaluate image quality

                         A number of datasets have been used to evaluate IQA algorithms, which have been
                         described in the literature. In this section, we describe some of the main datasets used
                         along with an indication of their size. Pires Dias et al. [3] used the public datasets of
                         DRIVE [42], Messidor [17], ROC [43] and STARE [44] and two proprietary datasets
                         (it is reported that access is available on request) from DR screening initiatives in the
                         central region of Portugal and Channai, India. The images from Portugal were graded
                         by a human grader from the Association for Innovation and Biomedical Research
                         on Light (AIBILI). An evaluation of the method was performed on a total of 2032
                         retinal images. Fleming et  al.  [33] used a private dataset of 1039 retinal images
                         from 489 patients attending a diabetic retinopathy screening program. The images
                         were graded by a clinician for image clarity and field of view. Hunter et al. [34],
                         used a private dataset of two hundred retinal fundus images randomly drawn from
                         a diabetic retinopathy screening dataset which were accompanied by a 1–5 scale
                         of image quality graded by an ophthalmologist, where low scores were associated
                         with poorer quality. Niemeijer et al. [36] used a total of 2000 images from a dataset
                         obtained from a diabetic retinopathy screening program in the Netherlands. Each
                         image was classified by an ophthalmologist to give an image quality category on a
                         scale from one to four. Giancardo et al. [37] used a dataset of 84 macula centered
                         fundus images. They were divided into four classes: “Good”, “Fair”, “Poor” and
                         “Outliers”. The first 3 classes were composed of a subset of a fundus images [11, 45].
                         Welikala et al. [6] utilized the UK Biobank dataset, which is a large scale population
                         based dataset, with access via an application process. The dataset included 68,151
                         participants. For evaluation of image quality 800 images were used from 400 random
                         participants. Images were defined by an experienced retinal grader as adequate or
                         inadequate. Paulus et al. [38] used a private dataset of 301 retinal images. Three
                         human observers, including one eye expert, decided on the quality of each image
                         according  to  set  criteria  and  they judged  the  image  quality  to  be  adequate  for  a
   143   144   145   146   147   148   149   150   151   152   153