Page 180 - Introduction to Autonomous Mobile Robots
P. 180

165
                           Perception
                           direction of illumination, the defocusing caused by optics, the side effects imposed by
                           nearby objects with different colors, and so on. Therefore the problem of visual feature
                           extraction is largely one of removing the majority of irrelevant information in an image so
                           that the remaining information unambiguously describes specific features in the environ-
                           ment.
                             We divide vision-based feature extraction methods into two classes based on their spa-
                           tial extent. Spatially localized features are those features found in subregions of one or
                           more images, corresponding to specific locations in the physical world. Whole-image fea-
                           tures are those features that are functions of the entire image or set of images, correspond-
                           ing to a large visually connected area in the physical world.
                             Before continuing it is important to note that all vision-based sensors supply images
                           with such a significant amount of noise that a first step usually consists of “cleaning” the
                           image before launching any feature extraction algorithm. Therefore, we first describe the
                           process of initial image filtering, or preprocessing.


                           Image preprocessing. Many image-processing algorithms make use of the second deriv-
                           ative of the image intensity. Indeed, the Laplacian of Gaussian method we studied in sec-
                           tion 4.1.8.2 for stereo ranging is such an example. Because of the susceptibility of such
                           high-order derivative algorithms to changes in illumination in the basic signal, it is impor-
                           tant to smooth the signal so that changes in intensity are due to real changes in the luminos-
                           ity of objects in the scene rather than random variations due to imaging noise. A standard
                           approach is convolution with a Gaussian distribution function, as we described earlier in
                           section 4.1.8.2:

                                ˆ   G ⊗
                                I =    I                                                     (4.82)
                             Of course, when approximated by a discrete kernel, such as a 3 x 3 table, the result is
                           essentially local, weighted averaging:


                                       121
                                     1
                                G =  ----- -  242                                            (4.83)
                                    16
                                       121

                             Such a low-pass filter effectively removes high-frequency noise, and this in turn causes
                           the first derivative and especially the second derivative of intensity to be far more stable.
                           Because of the importance of gradients and derivatives to image processing, such Gaussian
                           smoothing preprocessing is a popular first step of virtually all computer vision algorithms.
   175   176   177   178   179   180   181   182   183   184   185