抓住
计算机科学
人工智能
代表(政治)
卷积神经网络
推论
计算机视觉
领域(数学)
模式识别(心理学)
机器学习
数学
政治学
政治
程序设计语言
法学
纯数学
作者
Hu Cao,Guang Chen,Zhijun Li,Feng Qian,Jianjie Lin,Alois Knoll
出处
期刊:IEEE-ASME Transactions on Mechatronics
[Institute of Electrical and Electronics Engineers]
日期:2022-12-16
卷期号:28 (3): 1384-1394
被引量:35
标识
DOI:10.1109/tmech.2022.3224314
摘要
Deep learning methods have achieved excellent results in the field of grasp detection. However, deep learning-based models for general object detection lack the proper balance of accuracy and inference speed, resulting in poor performance in real-time grasp tasks. This work proposes an efficient grasp detection network with n-channel images as inputs for robotic grasp. The proposed network is a lightweight generative structure for grasp detection in one stage. Specifically, a Gaussian kernel-based grasp representation is introduced to encode the training samples, embodying the maximum center that possesses the highest grasp confidence. A receptive field block is plugged into the bottleneck to improve the model's feature discriminability. In addition, pixel-based and channel-based attention mechanisms are used to construct a multidimensional attention fusion network to fuse valuable semantic information, achieved by suppressing noisy features and highlighting object features. The proposed method is evaluated on the Cornell, Jacquard, and extended OCID grasp datasets. The experimental results show that our method achieves excellent balancing accuracy and running speed performance. The network gets a running speed of $\text{6}\,\text{ms}$ , achieving better performance on the Cornell, Jacquard, and extended OCID grasp datasets with 97.8, 95.6, and 76.4% accuracy, respectively. Subsequently, an excellent grasp success rate in a physical environment is obtained using the UR5 robot arm.
科研通智能强力驱动
Strongly Powered by AbleSci AI