计算机科学
渲染(计算机图形)
强化学习
人工智能
钥匙(锁)
机器人
功能(生物学)
动作(物理)
机器人学
人机交互
物理
计算机安全
量子力学
进化生物学
生物
作者
Shuran Song,Andy Zeng,Johnny Lee,Thomas Funkhouser
出处
期刊:IEEE robotics and automation letters
日期:2020-06-25
卷期号:5 (3): 4978-4985
被引量:131
标识
DOI:10.1109/lra.2020.3004787
摘要
Intelligent manipulation benefits from the capacity to flexibly control an end-effector with high degrees of freedom (DoF) and dynamically react to the environment. However, due to the challenges of collecting effective training data and learning efficiently, most grasping algorithms today are limited to top-down movements and open-loop execution. In this work, we propose a new low-cost hardware interface for collecting grasping demonstrations by people in diverse environments. This data makes it possible to train a robust end-to-end 6DoF closed-loop grasping model with reinforcement learning that transfers to real robots. A key aspect of our grasping model is that it uses “action-view” based rendering to simulate future states with respect to different possible actions. By evaluating these states using a learned value function (e.g., Q-function), our method is able to better select corresponding actions that maximize total rewards (i.e., grasping success). Our final grasping system is able to achieve reliable 6DoF closed-loop grasping of novel objects across various scene configurations, as well as in dynamic scenes with moving objects.
科研通智能强力驱动
Strongly Powered by AbleSci AI