Page 33 - Human Inspired Dexterity in Robotic Manipulation
P. 33

Sensorimotor Learning of Dexterous Manipulation  29


              include the hand’s greater number of degrees of freedom, and the fact that
              manipulation provides richer contextual cues than reaching (i.e., object
              geometry provides information about the object’s mass distribution), and
              therefore the way it would react to forces and torque exerted by the hand.
                 This chapter reviews recent studies and our current understanding of
              how humans learn to perform dexterous manipulation in the context of the-
              ories of sensorimotor learning and concludes with open questions and future
              directions for research on acquisition of dexterity in humans and robotic
              applications.

              3.2 LEARNING MANIPULATION: THEORETICAL
              AND EXPERIMENTAL EVIDENCE
              3.2.1 Background

              Sensorimotor learning and memory have been studied extensively in motor
              neuroscience using adaptation experiments in reaching and oculomotor
              tasks [9]. In general, all skilled motor behavior relies on learning both control
              and prediction which can be considered as two reciprocal procedures: control
              generates motor commands to produce desired consequences, whereas pre-
              diction maps motor commands (i.e., efferent copy) into expected sensory
              consequences [10]. The mechanisms underlying these two procedures have
              been traditionally proposed as inverse and forward internal models, respectively
              [4]. The update of internal models is considered to be driven by the error
              caused by a mismatch between sensed and predicted sensory outcome on
              a trial-by-trial basis. This error is assumed to signal the direction of adjust-
              ment for the next trial for error reduction. Error-based learning has been
              demonstrated in many motor adaptation schemes, including saccade adap-
              tation [11], visuomotor adaptation [12], and force-field adaptation [7].
              State-space models are often used to describe error-based updates by defin-
              ing the estimation of task dynamics as states that can be modified by errors
              with learning rates. The exact structure of the model may vary as it could
              consist of a single learning rate [13] or multiple rates [5], and could include
              context selectors [14]. It has been demonstrated that these models can
              capture many well-known phenomena including interference and savings.
                 Recent research has also demonstrated other sensorimotor learning pro-
              cesses. Besides explaining additional performance improvements after errors
              have been minimized, these nonerror based learning processes were also
              found to be operating in parallel with error-based learning [6]. Reinforcement
              learning uses signals, such as success and failure, which do not provide the
   28   29   30   31   32   33   34   35   36   37   38