DiPGrasp: Parallel Local Searching for Efficient Differentiable Grasp Planning

Supplementary File

Supplementary File for detailed Application setting and more experiments: link 

Abstract

Grasp planning is an important task for robotic manipulation. Though it is a richly studied area, a standalone, fast, and differentiable grasp planner that can work with robot grippers of different DOFs has not been reported. In this work, we present DiPGrasp, a grasp planner to satisfy all these goals. DiPGrasp takes a geometric surface matching grasp quality metric. It adopts a gradient-based optimization scheme on the metric which also considers parallel sampling and collision handling. This not only drastically accelerates the grasp search process over the object surface but also makes it differentiable. We apply DiPGrasp to three applications, namely grasp dataset construction, mask-conditioned planning, and pose refinement. For dataset generation, as a standalone planner, DiPGrasp has clear advantages over speed and quality in comparison with several classic planners. For mask-conditioned planning, it can turn a 3D perception model into a 3D grasp detection model instantly. As a pose refiner, it can optimize the coarse grasp prediction from the neural network, as well as the neural network parameters. Finally, we conduct real-world experiments with the Barrett hand and Schunk SVH 5-finger hand.

Method

DiPGrasp  takes  a  point  cloud  with  normal  as  input.  It  first  samples  locations  on  the  point  cloud  (red  dot)  and  initializes  the  pose

accordingly. Then it operates the differentiable optimization process to generate the grasps.

Applications

Grasp Dataset Generation 

To make the grasp evaluation in simulation can be conducted in parallel.
1. We load object and the corresponding grasp pose, the initial object pose is adjust to hand-centered pose.

2. The gripper will try to grasp the object, and lift it up 20cm.

3. We adjust the gravity directions multiple times.

If the object is still in gripper's hand, we consider it a valid grasp after physics-based simulation evaluation.

Scene Construction

Finally, we put the object models into a scene, and generate the IR-based scene point cloud, 3D annotations and grasp annotations for later use in Mask-conditioned planning.

Mask-Conditioned Planning

Pose Refinement

Real-World Experiments