Page 254 - Innovations in Intelligent Machines
P. 254

Toward Robot Perception through Omnidirectional Vision  247
                           scattered throughout a high dimensional space but – due to their similarity –
                           lie in a lower dimensional subspace.
                              We implement dimensionality reduction using the classical procedure of
                                                            8
                           Principal Component Analysis (PCA) , as described by Murase and Nayar in
                           [62], and detailed by Winters in [88] or Gaspar, Winters and Santos-Victor in
                           [35]. Simply put, Principal Component Analysis reduces the dimension-
                           ality of a set of linearly independent input variables, while still accurately
                           representing most of the original data. The reconstruction of this original
                           data is optimal in the sense that the mean square error between it and the
                           original data is minimized.
                                                                                      L
                              Imagine that we represent images as L-dimensional vectors in R .Dueto
                           the similarity between images (data redundancy) these vectors will not span
                                             L
                           the entire space of R but rather, they will be confined (or close, to be more
                           precise) to a lower-dimensional subspace, R M  where M<< L. Hence, to save
                           on computation, we can represent our images by their co-ordinates in such
                           a lower-dimensional subspace, rather than using all of the pixel information.
                           Each time a new image is acquired, its capture position can easily be deter-
                           mined by projecting it into the lower-dimensional subspace and finding its
                           closest match from the a priori set of points (images).
                              A basis for such a linear subspace can be found through PCA, where
                           the basis vectors are denominated Principal Components. They can be
                           computed as the eigenvectors of the covariance matrix of the normalised
                           set of images acquired by the robot. The number of eigenvectors that can be
                           computed in such a way is the same as the number of images in the input
                           data, and the eigenvectors are the same size as the images.
                              Each reference image is associated with a qualitative robot position (e.g.
                           half way along the corridor). To find the robot position in the topological
                           map, we have to determine the reference image that best matches the current
                           view. The distance between the current view and the reference images can be
                           computed directly using their projections (vectors) on the lower dimensional
                           eigenspace. The distance is computed between M-dimensional coefficient vec-
                           tors (typically 10 to 12), as opposed to image size vectors (128 × 128). The
                           position of the robot is that associated with the reference image having the
                           lowest distance.
                              When using intensity images, comparison of images is essentially a sum
                           of squared differences of brightness (radiance) values and because of this the
                           robot is prone to miscalculating its location where large non-uniform devi-
                           ations in illumination occur. This can be overcome by using edge images,
                           although these are not robust to edge-point position errors. The solution
                           therefore, is to compare shapes instead of edge-points, or more specifically
                           the distance between shapes present in both images. There are several possible



                            8
                             It is sometimes known as the application of the Karhunen-Lo`eve transform
                             [67, 86].
   249   250   251   252   253   254   255   256   257   258   259