Page 47 - Dynamic Vision for Perception and Control of Motion
P. 47

2.1 Three-dimensional (3-D) Space and Time      31



              The challenge in machine vision as opposed to computer graphics is that some
            of the transformation parameters entering the matrices are not known beforehand
            but are the unknowns of the vision process, which have to be determined from im-
            age sequence analysis. Therefore, in each transformation, its sensitivity to small
            parameter changes has to be determined to compute the corresponding overall
            “Jacobian” matrices (JM, the first-order approximation for the nonlinear functional
            relationship describing the mapping of features on objects in the real world to those
            measured in the images). This rather compute-intensive operation and an efficient
            implementation will be discussed in Section 2.1.2.
              The tendency toward separation  of application-oriented aspects from those
            geared to the general methods of dynamic vision required a major change from the
            initial approach with respect to handling homogeneous coordinates. Concatenation
            is shifted to the evaluation of the scene model at runtime; then, both the nominal
            total HTM and the partial-derivative matrices for all unknown parameters and state
            variables are computed in conjunction (maybe numerically). This allows efficient
            use of intermediate results and makes the setup of new problems much easier for
            the user. The corresponding representation scheme for all objects and CSs in a so-
            called “scene  tree” has been developed by  D. Dickmanns (1997) and  will be dis-
            cussed in the following paragraphs.
              Figure 2.7 without the shaded areas gives an example of a scene tree for describ-
            ing the geometrical relations among several objects of relevance for the vision task
                                                      shown in  Figure  2.6 a single
                 [3 (known) translations],  Vehicle body  camera on a straight road. The
                    2 rotations  \ , 4 cb             nodes and edges in the shaded
                           cb
                                       3 translations,
                    Camera             3 rotations    areas  on the right-hand  side
                                       (general case)
              1 translation y gc ,                    and on top will be needed for
             2 rotations  \ , 4                       the more general case  of  a
                    c  c                road nearby
                              range &                 camera onboard a vehicle
                 Perspective  bearing    Curvature
                 projection         (a)  parameters   moving on a  curved road. In
                 Chip                    C N0 , C N1 (L N )  the straight  road scene, the
                               object
            Frame grabber                 Road at ob-  “object” represents the rod on
                    Pixel          (b)   ject location  the road at some  look-ahead
                    position
                           1 translation y go  Curvature  distance x co; its lateral position
                           1 rotation  \ o  parameters
                                           C F0 , C F1 (L F )  on the road can  be  recovered
                   Image in
                   storage                            in the image from  the road
                   location      ‘Vanishing  Road     boundaries nearby  and from
                                   point’  far away
                              for straight road
                                                      the vanishing point at the hori-
                                                      zon (see Figure 2.8).The figure
            Figure 2.7. Scene tree for representing spatial rela-
                                                      shows the resulting image, into
            tionships between objects seen and their image in
                                                      which some  labels for later
            perspective projection
                                                      image interpretation have been
                                                      inserted.
              For a horizontal straight road with parallel lines, the vanishing point, at which
            all parallel lines intersect, lies on the horizon line. Its distance components to the
            image center yield the direction of the optical axis: í\ c to the direction of the road
            and íT c to the horizon line. The center of gravity (cg) of the rod has the averaged
            coordinates of the end points E and T in 3-D space.
   42   43   44   45   46   47   48   49   50   51   52