Abstract | ||
---|---|---|
Physical contact events often allow a natural decomposition of manipulation tasks into action phases and subgoals. Within the motion primitive paradigm, each action phase corresponds to a motion primitive, and the subgoals correspond to the goal parameters of these primitives. Current state-of-the-art reinforcement learning algorithms are able to efficiently and robustly optimize the parameters of motion primitives in very high-dimensional problems. These algorithms often consider only shape parameters, which determine the trajectory between the start- and end-point of the movement. In manipulation, however, it is also crucial to optimize the goal parameters, which represent the subgoals between the motion primitives. We therefore extend the policy improvement with path integrals (PI$^2$) algorithm to simultaneously optimize shape and goal parameters. Applying simultaneous shape and goal learning to sequences of motion primitives leads to the novel algorithm PI$^2$ Seq. We use our methods to address a fundamental challenge in manipulation: improving the robustness of everyday pick-and-place tasks. |
Year | DOI | Venue |
---|---|---|
2012 | 10.1109/TRO.2012.2210294 | IEEE Transactions on Robotics |
Keywords | Field | DocType |
Learning systems,Adaptive systems,Learning,Manipulators,Grasping | Path integral formulation,Motion control,Computer science,Control engineering,Robustness (computer science),Theoretical computer science,Artificial intelligence,Trajectory control,Trajectory,Reinforcement learning | Journal |
Volume | Issue | ISSN |
28 | 6 | 1552-3098 |
Citations | PageRank | References |
19 | 0.81 | 10 |
Authors | ||
3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Freek Stulp | 1 | 448 | 40.02 |
Evangelos A. Theodorou | 2 | 807 | 70.91 |
Stefan Schaal | 3 | 6081 | 530.10 |