Learning Optimal Manipulation Skills in a Human-like Way for Contact-rich Tasks

Abstract


Imitation Learning(IL) algorithms have shown great success in human-robot manipulation skills transferring. However, they assume access to optimal human demonstrations and don't explicitly model the environment dynamics, which limits the optimality of learned manipulation skills and their performance in solving contact-rich tasks. Humans learn professional manipulation skills through imitating experts and improving the learned skills from trials and errors. In this letter, we develop a human-like safe optimal manipulation skills learning framework for solving contact-rich tasks. This framework combines IL methods with Reinforcement Learning (RL) algorithms for learning initial manipulation skills by imitating human demonstrations and further optimizing the initial skills with trials and errors. Furthermore, we propose using demonstration space to constrain the learned optimal actions in continuous action space for the safe exploration of RL agents. We also further develop our learning framework to learn optimal variable impedance manipulation skills for avoiding large contact forces and ensuring interaction safety. Simulation and real-world experiments on a 7 DoF redundant robot manipulator for peg insertion tasks validate the effectiveness of our proposed method.

Overview:


Overview of the proposed method.