Visual Manipulation Relationship Network for Autonomous Robotics

2018 IEEE-RAS 18th International Conference on Humanoid Robots (Humanoids)(2018)

引用 47|浏览79
暂无评分
摘要
Robotic grasping is one of the most important fields in robotics, in which great progress has been made in recent years with the help of convolutional neural network (CNN). However, including multiple objects in one scene can invalidate the existing CNN-based grasp detection algorithms, because manipulation relationships among objects are not considered, which are required to guide the robot to grasp things in the right order. This paper presents a new CNN architecture called Visual Manipulation Relationship Network (VMRN) to help robots detect targets and predict the manipulation relationships in real time, which ensures that the robot can complete tasks in a safe and reliable way. To implement end-to-end training and meet real-time requirements in robot tasks, we propose the Object Pairing Pooling Layer (OP 2 L) to help to predict all manipulation relationships in one forward process. Moreover, in order to train VMRN, we collect a dataset named Visual Manipulation Relationship Dataset (VMRD) consisting of 5185 images with more than 17000 object instances and the manipulation relationships between all possible pairs of objects in every image, which is labeled by the manipulation relationship tree. The experimental results show that the new network architecture can detect objects and predict manipulation relationships simultaneously and meet the real-time requirements in robot tasks.
更多
查看译文
关键词
autonomous robotics,robotic grasping,convolutional neural network,manipulation relationships,real-time requirements,robot tasks,manipulation relationship tree,visual manipulation relationship dataset,visual manipulation relationship network,CNN-based grasp detection algorithms,CNN architecture,object pairing pooling layer,network architecture,object detection,target detection
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要