Page 156 - Biomedical Engineering and Design Handbook Volume 1, Fundamentals
P. 156

BIOMECHANICS OF HUMAN MOVEMENT  133

                          depending on the ambient lighting, but have been turned
                          off for the sake of this example. Figure 6.5 is also help-
                          ful in depicting the effect of using binary thresholding to
                          suppress background light sources or reflective objects
                          other than the tracking targets. The idea is to suppress all
                          voltages below a user-specified threshold, which has the
                          effect of moving the light-reflective tracking targets to
                          the foreground.  This is ideal since we are ultimately
                          concerned with locating the center of a prospective
                          target in u, v coordinates and do not want other sources
                          of light affecting the location of the computed center.
                          One method of determining the center of a target is to  FIGURE 6.5 Schematic representation of the
                                                                     light-sensitive elements of the imaging sensor.
                          scan the matrix of light sensitive elements for transitions  The silver circle is the projection of a target onto
                          in voltage (i.e., edge detection) and fit a circle of best fit  the image plane. Note the direction of the u, v
                          to the resulting “edges.” While this approach is concep-  reference axes.
                          tually straight forward, binary thresholding as described
                          here would eliminate useful information that could otherwise be used to determine the center of the
                          projected target at greater subpixel accuracy. For example, rather than simply treating each element
                          of the sensor as being on or off, we could use a weighted average of sensor element voltages and
                          a geometric constraint that the active elements form a circle. The center of the target in the image
                          plane is assumed to lie at the center of the circle.
                            If one were to draw a line from the center of the target in the image plane to the X, Y, Z coordi-
                          nates of the target in object-space, it would be clear that the mapping between these spaces is not
                          unique since all targets lying on this line would map to the same u, v coordinates. From this it is evi-
                          dent that the location of a target in object-space cannot be determined using only one camera. This
                          raises a subtle but important distinction regarding the role of the camera in video-based motion
                          analysis. The camera does not actually record the location of a target in object-space, but rather
                          the role of the camera is to define a ray in the direction of the target. When multiple cameras view
                          the same target, the location of the target in object-space is assumed to lie at the intersection of the
                          directed rays from each camera. The cameras must first be calibrated before the intersection of these
                          rays can be calculated.

                          Camera Calibration.  Each camera must be calibrated before it can contribute to locating a target
                          in object-space. Camera calibration defines a mapping from three-dimensional object-space into the
                          two-dimensional u, v coordinates of the camera. This mapping is expressed in Eq. (6.2) using homo-
                          geneous coordinates:

                                                                ⎛  X ⎞
                                                         ⎛ λu⎞  ⎜ ⎟
                                                                Y
                                                         ⎜  λv = A ⎜ ⎟
                                                           ⎟
                                                         ⎜  ⎟   ⎜  Z ⎟                        (6.2)
                                                         ⎝  λ ⎠  ⎜ ⎟
                                                                ⎝ ⎠
                                                                 1
                          where λ is a scale factor relating the spaces, u and v are the image plane coordinates of a target,
                          A is a 3  × 4 transformation matrix, and  X, Y,  Z are the coordinates of a target in object-space.
                          Expanding the right-hand side of Eq. (6.2) results in the following set of equations:
                                                  λu =  α X + α Y + α Z + α 14                (6.3)
                                                       11
                                                             12
                                                                  13
                                                   λv =  α X + α Y +  α Z + α  24             (6.4)
                                                                  23
                                                       21
                                                             22
                                                   λ =  α X  +  α Y  +  α Z  +  α 34          (6.5)
                                                             32
                                                       31
                                                                  33
   151   152   153   154   155   156   157   158   159   160   161