Learning to Dexterously Pick or Separate Tangled-Prone Objects for Industrial Bin Picking

Xinyi Zhang1, Yukiyasu Domae2, Weiwei Wan1, Kensuke Harada1,2

1Osaka University, 2National Institute of Advanced Industrial Science and Technology (AIST)

Industrial bin picking for tangled-prone objects requires the robot to either pick up untangled objects or perform separation manipulation when the bin contains no isolated objects. The robot must be able to flexibly perform appropriate actions based on the current observation. It is challenging due to high occlusion in the clutter, elusive entanglement phenomena, and the need for skilled manipulation planning. In this paper, we propose an autonomous, effective and general approach for picking up tangled-prone objects for industrial bin picking. First, we learn PickNet - a network that maps the visual observation to pixel-wise possibilities of picking isolated objects or separating tangled objects and infers the corresponding grasp. Then, we propose two effective separation strategies: Dropping the entangled objects into a buffer bin to reduce the degree of entanglement; Pulling to separate the entangled objects in the buffer bin planned by PullNet - a network that predicts position and direction for pulling from visual input. To efficiently collect data for training PickNet and PullNet, we embrace the self-supervised learning paradigm using an algorithmic supervisor in a physics simulator. Real-world experiments show that our policy can dexterously pick up tangled-prone objects with success rates of 90%. We further demonstrate the generalization of our policy by picking a set of unseen objects.

Paper

Latest version: here (with supplementary material)

IEEE Robotics and Automation Letters (RA-L), 2023.

Code is available on Github.

Supplementary Video

Results of predicting isolated/tangled objects using PickNet

PickNet outputs two heatmaps: PickMap and SepMap, respectively indicate the grasp affordance of isolated or tangled objects in the clutter.

Results of planning pulling actions using PullNet

We rotate the input depth images for 8 angles representing 8 directions of pulling (pointing to the right). We select the one with the highest values among the PullNet outputs (PullMaps).

Data collection in simulation

Simulated picking demonstration using various objects

The following videos show the demonstrations for data collection in a phyics simulator.

The following figures shows the ground-truth of PickNet/PullNet data.

Separation strategy 1: Dropping entangled objects in a buffer bin to separate.

The following videos show the robot dropping the entangled objects in the buffer bin.

Seen objects

Unseen objects

Separation stragegy 2: Pulling to disentangle various objects.

The following videos show the robot untangles the objects by pulling combined with wiggling.

Seen objects

Unseen objects

Video: PullNet can plan pulling actions for various patterns of entanglement.

We test PullNet using different patterns of the entanglement in the buffer bin.

Seen object

Unseen object

Failure modes and limitations

Grasp failure

Challenging entanglmenet patterns

Unsuitable object shapes

Acknowledgements

This work is supported by Toyota Motor Coorporation.