强化学习
避障
计算机科学
模块化设计
任务(项目管理)
障碍物
一般化
人工智能
建筑
网络体系结构
实时计算
人机交互
分布式计算
机器人
工程类
移动机器人
计算机网络
艺术
数学分析
数学
系统工程
法学
政治学
视觉艺术
操作系统
作者
Yuanda Wang,Haibo He,Changyin Sun
标识
DOI:10.1109/tg.2018.2849942
摘要
In this paper, we propose an end-to-end modular reinforcement learning architecture for a navigation task in complex dynamic environments with rapidly moving obstacles. In this architecture, the main task is divided into two subtasks: local obstacle avoidance and global navigation. For obstacle avoidance, we develop a two-stream Q-network, which processes spatial and temporal information separately and generates action values. The global navigation subtask is resolved by a conventional Q-network framework. An online learning network and an action scheduler are introduced to first combine two pretrained policies, and then continue exploring and optimizing until a stable policy is obtained. The two-stream Q-network obtains better performance than the conventional deep Q-learning approach in the obstacle avoidance subtask. Experiments on the main task demonstrate that the proposed architecture can efficiently avoid moving obstacles and complete the navigation task at a high success rate. The modular architecture enables parallel training and also demonstrates good generalization capability in different environments.
科研通智能强力驱动
Strongly Powered by AbleSci AI