Publications Freek Stulp


Back to Homepage
Sorted by DateClassified by Publication TypeClassified by Research Category
Reinforcement Learning of Impedance Control in Stochastic Force Fields
Freek Stulp, Jonas Buchli, Alice Ellmer, Michael Mistry, Evangelos Theodorou, and Stefan Schaal. Reinforcement Learning of Impedance Control in Stochastic Force Fields. In International Conference on Development and Learning (ICDL), 2011.
Download
[PDF]1.1MB  
Abstract
Variable impedance control is essential for ensuring robust and safe physical interaction with the environment. As demonstrated in numerous force field experiments, humans combine two strategies to adapt their impedance to external perturbations: 1) if perturbations are unpredictable, subjects increase their impedance through co-contraction; 2) if perturbations are predictable, subjects learn a feed-forward command to counter the known perturbation. In this paper, we apply the force field paradigm to a simulated 7-DOF robot, by exerting stochastic forces on the robot's end-effector. The robot `subject' uses our model-free reinforcement learning algorithm PI2 to simultaneously learn the end-effector trajectories and variable impedance schedules. We demonstrate how the robot learns the same two-fold strategy to perturbation rejection as humans do, resulting in qualitatively similar behavior. Our results provide a biologically plausible approach to learning appropriate impedances purely from experience, without requiring a model of either body or environment dynamics.
BibTeX
@InProceedings{stulp11reinforcement,
  title                    = {Reinforcement Learning of Impedance Control in Stochastic Force Fields},
  author                   = {Freek Stulp and Jonas Buchli and Alice Ellmer and Michael Mistry and Evangelos Theodorou and Stefan Schaal},
  booktitle                = {International Conference on Development and Learning (ICDL)},
  year                     = {2011},
  abstract                 = {Variable impedance control is essential for ensuring robust and safe physical interaction with the environment. As demonstrated in numerous force field experiments, humans combine two strategies to adapt their impedance to external perturbations: 1)~if perturbations are unpredictable, subjects increase their impedance through co-contraction; 2)~if perturbations are predictable, subjects learn a feed-forward command to counter the known perturbation. In this paper, we apply the force field paradigm to a simulated {7-DOF} robot, by exerting stochastic forces on the robot's end-effector. The robot `subject' uses our model-free reinforcement learning algorithm PI2 to simultaneously learn the end-effector trajectories and variable impedance schedules. We demonstrate how the robot learns the same two-fold strategy to perturbation rejection as humans do, resulting in qualitatively similar behavior. Our results provide a biologically plausible approach to learning appropriate impedances purely from experience, without requiring a model of either body or environment dynamics.},
  bib2html_accrate         = {Oral: 26\%},
  bib2html_pubtype         = {Refereed Conference Paper},
  bib2html_rescat          = {Reinforcement Learning of Variable Impedance Control}
}

This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints.


Generated by bib2html.pl (written by Patrick Riley ) on Mon Jul 20, 2015 21:50:11