Embodied Uncertainty-Aware Object Segmentation
Xiaolin Fang, Leslie Pack Kaelbling, Tomás Lozano-Pérez
MIT CSAIL
{xiaolinf,lpk,tlp}@csail.mit.edu
Xiaolin Fang, Leslie Pack Kaelbling, Tomás Lozano-Pérez
MIT CSAIL
{xiaolinf,lpk,tlp}@csail.mit.edu
We introduce uncertainty-aware object instance segmentation (UncOS) and demonstrate its usefulness for embodied interactive segmentation. To deal with uncertainty in robot perception, we propose a method for generating a hypothesis distribution of object segmentation. We obtain a set of region-factored segmentation hypotheses together with confidence estimates by making multiple queries of large pre-trained models. This process can produce segmentation results that achieve state-of-the-art performance on unseen object segmentation problems. The output can also serve as input to a belief-driven process for selecting robot actions to perturb the scene to reduce ambiguity. We demonstrate the effectiveness of this method in real-robot experiments.
UncOS takes an RGB (+Depth optionally) image and generates multiple segmentation hypotheses of the scene and their associated confidence.
To use it as an unseen object instance segmentation (UOIS) method, we take the most likely/confident hypothesis reported by UncOS.
The results shown below are the most confident hypothesis reported by UncOS on the OCID testset. These are NOT carefully selected (uniformly sampled from the test set).
Click on the images to expand and see more results.
We use a robot to selectively poke objects in the scene using a simple greedy strategy that attempts to select a small perturbation that will maximize information gain.
Regions correctly segmented after embodied interaction are highlighted with green circles (errors highlighted with red circles).