Page 143 - Introduction to Autonomous Mobile Robots
P. 143

128


                                                      1
                                                                  2
                                                                           2
                                                     --------- if ( x –  x ) + ( y – y ) ) ≤  R 2  Chapter 4
                                                           (
                                hx y x y R,(  ,  ,  ,  ) =  2  g  f   g   f                  (4.23)
                                               ,
                                   g  g  f  f  xy    πR
                                                                           2
                                                                  2
                                                      0 if ((  x –  x ) + ( y – y ) ) >  R 2
                                                                      g
                                                                          f
                                                             g
                                                                 f
                             Intuitively, point   contributes to the image pixel  x y,(  g  g )   only when the blur circle of
                                           P
                           point   contains the point  x y,(  g  g ) . Now we can write the general formula that computes
                                P
                           the value of each pixel in the image, f x y,(  )  , as a function of the point spread function and
                           the focused image:
                                                ,
                                                  ,,,
                                gx y,(  ) =  ∑ hx y xyR   )fxy)                              (4.24)
                                                             ,
                                             (
                                                            (
                                                         ,
                                   g  g        g  g     xy
                                           ,
                                          xy
                                                                                         g
                                                                     R
                             This equation relates the depth of scene points via   to the observed image  . Solving
                              R
                           for   would provide us with the depth map. However, this function has another unknown,
                           and that is  , the focused image. Therefore, one image alone is insufficient to solve the
                                    f
                           depth recovery problem, assuming we do not know how the fully focused image would
                           look.
                             Given two images of the same scene, taken with varying camera geometry, in theory it
                                                 g
                                                                     f
                           will be possible to solve for   as well as   because   stays constant. There are a number
                                                            R
                           of algorithms for implementing such a solution accurately and quickly. The classic
                           approach is known as inverse filtering because it attempts to directly solve for  R  , then
                           extract depth information from this solution. One special case of the inverse filtering solu-
                           tion has been demonstrated with a real sensor. Suppose that the incoming light is split and
                           sent to two cameras, one with a large aperture and the other with a pinhole aperture [121].
                           The pinhole aperture results in a fully focused image, directly providing the value of  . f
                           With this approach, there remains a single equation with a single unknown, and so the solu-
                           tion is straightforward. Pentland [121] has demonstrated such a sensor, with several meters
                           of range and better than 97% accuracy. Note, however, that the pinhole aperture necessi-
                           tates a large amount of incoming light, and that furthermore the actual image intensities
                           must be normalized so that the pinhole and large-diameter images have equivalent total
                           radiosity. More recent depth from defocus methods use statistical techniques and charac-
                           terization of the problem as a set of linear equations [64]. These matrix-based methods have
                           recently achieved significant improvements in accuracy over all previous work.
                             In summary, the basic advantage of the depth from defocus method is its extremely fast
                           speed. The equations above do not require search algorithms to find the solution, as would
                           the correlation problem faced by depth from stereo methods. Perhaps more importantly, the
                           depth from defocus methods also need not capture the scene at different perspectives, and
                           are therefore unaffected by occlusions and the disappearance of objects in a second view.
   138   139   140   141   142   143   144   145   146   147   148