Teng Xue, Amirreza Razmjoo, Suhan Shetty, and Sylvain Calinon
Conference on Robot Learning (CoRL), 2024
Contact-rich manipulation plays an important role in everyday life, but uncertain parameters pose significant challenges to model-based planning and control. To address this issue, domain adaptation and domain randomization have been proposed to learn robust policies. However, they either lose the generalization ability to diverse instances or perform conservatively due to neglecting instance-specific information. In this paper, we propose a bi-level approach to learn robust manipulation primitives, including parameter-augmented policy learning using multiple models with tensor approximation, and parameter-conditioned policy retrieval through domain contraction. This approach unifies domain randomization and domain adaptation, providing optimal behaviors while keeping generalization ability. We validate the proposed method on three contact-rich manipulation primitives: hitting, pushing, and reorientation. The experimental results showcase the superior performance of our approach in generating robust policies for instances with diverse physical parameters.
Overview of the proposed bi-level approach. Left: Parameter-augmented policy training using multiple models. The state, action, and parameter variables are denoted in black, blue, and red colors, respectively. Right: Parameter-conditioned policy retrieval through domain contraction. The retrieved policies perform well in terms of both generalization and optimality given a diverse set of objects with different shapes, weights, and friction parameters.
The details of policy learning and retrieval are shown as below:
Contact: teng.xue@idiap.ch
Feel free to contact if you have any questions!