基于单视图6维位姿估计的机器人抓取
Robotic Grasping Based on 6D Pose Estimation from Single View
-
摘要: 现实场景中物体种类多样、摆放位置随机,会导致智能机器人物体识别困难,抓取成功率不高。针对这一问题,提出一种在遮挡、同类多目标、堆叠等复杂情况下机器人抓取的方法。基于通道注意力机制ECA (Efficient Channel Attention)和残差网络ResNet (Residual Network),设计了编解码器结构的单视图6维位姿估计网络;利用合成数据集制作方法生成了6维位姿估计和抓取训练数据集;机器人抓取控制模块根据6维位姿估计网络的输出以及手眼标定结果,控制UR5机器人实现智能抓取。在Linemod、YCB-Video以及本文合成数据集上的实验结果表明,所提方法的平均抓取成功率达到95%。Abstract: In real-world scenarios, the diversity of object types and random placement can lead to difficulties in object recognition for intelligent robots, resulting in a low success rate in grasping. A method for robot grasping in complex situations such as occlusion, multiple targets of the same type, and stacking is proposed to address this issue. A single view 6D pose estimation network with encoder decoder structure is designed based on channel attention mechanism ECA and residual network ResNet; A 6D pose estimation and grasping training dataset is generated using a synthetic dataset production method; The robot grasping control module controls the UR5 robot to achieve intelligent grasping based on the output of the 6D pose estimation network and the results of hand eye calibration. The experimental results on Linemod, YCB-Video, and the synthesized dataset show that the average grasping success rate of our method reaches 95%.