Under Review
Applying sim-to-real RL for Deformable and Fragile Object Manipulation remains largely unexplored. We propose the first visuomotor learning framework for sim-to-real manipulation of 3D deformable and fragile objects.
Picking up tofu trained with a vanila RL policy often leads to permanent damage.
We learn gentle and safe manipulation policies using object internal stress, obtained in simulation, as a stress-penalized reward.
We bootstrap it with a rigid-to-soft curriculum and human demonstrations to encourage policy convergence and increase training speed.
Learning to gently pick up a tofu block with significantly reduced stress.
(1) A vanilla RL policy generates a high stress spike:
(2) Our approach ensures the stress applied to the tofu remains low:
Our policy can be transferred from simulation to the real world zero-shot.
(1) Vanilla RL baseline: tofu is significantly damaged.
(2) Our method: tofu remains intact.
(1) Vanilla RL baseline: tofu is significantly damaged and goal is not reached.
(2) Our method: tofu remains intact and goal is reached.
We replace the previous cylindrical tofu by cubic tofu.
(1) Vanilla RL baseline: tofu is significantly damaged.
(2) Our method: tofu remains intact and goal is reached.