抓住
计算机科学
人工智能
可视化
任务(项目管理)
边距(机器学习)
过程(计算)
融合机制
融合
机器人
计算机视觉
机器学习
人机交互
工程类
哲学
操作系统
脂质双层融合
程序设计语言
系统工程
语言学
作者
Shaowei Cui,Rui Wang,Junhang Wei,Jingyi Hu,Shuo Wang
出处
期刊:IEEE robotics and automation letters
日期:2020-07-21
卷期号:5 (4): 5827-5834
被引量:50
标识
DOI:10.1109/lra.2020.3010720
摘要
Predicting whether a particular grasp will succeed is critical to performing stable grasping and manipulating tasks. Robots need to combine vision and touch as humans do to accomplish this prediction. The primary problem to be solved in this process is how to learn effective visual-tactile fusion features. In this letter, we propose a novel Visual-Tactile Fusion learning method based on the Self-Attention mechanism (VTFSA) to address this problem. We compare the proposed method with the traditional methods on two public multimodal grasping datasets, and the experimental results show that the VTFSA model outperforms traditional methods by a margin of 5+% and 7+%. Furthermore, visualization analysis indicates that the VTFSA model can further capture some position-related visual-tactile fusion features that are beneficial to this task and is more robust than traditional methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI