Robotic Grasp Detection Using Structure Prior Attention and Multiscale Features

Lu Chen, Mingdi Niu, Jing Yang, Yuhua Qian, Zhuomao Li, Keqi Wang, Tao Yan, Panfeng Huang

科研成果: 期刊稿件文章同行评审

5 引用 (Scopus)

摘要

Most available grasp detection methods tend to directly predict grasp configurations with deep neural networks, where all features are equally extracted and utilized, leading to the relative restriction of truly useful grasping features. Inspired by the observed three-section structure pattern revealed by human-labeled graspable rectangles, we first design a structure prior attention (SPA) module which uses two-dimensional encoding to enhance the local patterns and utilizes self-attention mechanism to reallocate distribution of grasping-specific features. Then, the proposed SPA module is integrated with fundamental feature extraction modules and residual connection to achieve the implicit and explicit feature fusion, which further serves as the building block of our proposed Unet-like grasp detection network. It takes RGBD images as input and outputs image-size feature maps, from which the grasp configurations can be determined. Extensive comparative experiments on the five public datasets prove our method's superiority to other approaches in detection accuracy, achieving 99.2%, 96.1%, 98.0%, 86.7%, and 92.6% on the Cornell, Jacquard, Clutter, VMRD, and GraspNet datasets. With visual evaluation metrics and user study, the quality maps generated by our method possess more concentrative distribution of high-confidence grasps and clearer discrimination with backgrounds. In addition, its effectiveness is also verified by robotic grasping under real-world scenario, leading to higher success rate.

源语言英语
页(从-至)7039-7053
页数15
期刊IEEE Transactions on Systems, Man, and Cybernetics: Systems
54
11
DOI
出版状态已出版 - 2024

指纹

探究 'Robotic Grasp Detection Using Structure Prior Attention and Multiscale Features' 的科研主题。它们共同构成独一无二的指纹。

引用此