We develop a system that enables remote teleoperation for 6-DoF robot control and a natural human intervention mechanism well suited to robot manipulation.
We introduce Intervention Weighted Regression (IWR), a simple yet effective method to learn from human interventions that encourages the policy to learn how to traverse bottlenecks through the interventions.
We evaluate our system and method on two challenging contact-rich manipulation tasks: a threading task and coffee machine task. We demonstrate that (1) policies trained on data collected by our system outperform policies trained on an equivalent amount of full human demonstration trajectories, (2) IWR outperforms alternatives for learning from the intervention data, and (3) our results hold across data collected from multiple human operators.