强化学习
计算机科学
避障
运动规划
人工智能
趋同(经济学)
路径(计算)
算法
适应性
一般化
机器人
移动机器人
数学
生态学
数学分析
程序设计语言
经济
生物
经济增长
作者
Hui Hu,Yuge Wang,Wenjie Tong,Jiao Zhao,Yulei Gu
出处
期刊:Applied sciences
[Multidisciplinary Digital Publishing Institute]
日期:2023-09-06
卷期号:13 (18): 10056-10056
被引量:14
摘要
Autonomous vehicles can reduce labor power during cargo transportation, and then improve transportation efficiency, for example, the automated guided vehicle (AGV) in the warehouse can improve the operation efficiency. To overcome the limitations of traditional path planning algorithms in unknown environments, such as reliance on high-precision maps, lack of generalization ability, and obstacle avoidance capability, this study focuses on investigating the Deep Q-Network and its derivative algorithm to enhance network and algorithm structures. A new algorithm named APF-D3QNPER is proposed, which combines the action output method of artificial potential field (APF) with the Dueling Double Deep Q Network algorithm, and experience sample rewards are considered in the experience playback portion of the traditional Deep Reinforcement Learning (DRL) algorithm, which enhances the convergence ability of the traditional DRL algorithm. A long short-term memory (LSTM) network is added to the state feature extraction network part to improve its adaptability in unknown environments and enhance its spatiotemporal sensitivity to the environment. The APF-D3QNPER algorithm is compared with mainstream deep reinforcement learning algorithms and traditional path planning algorithms using a robot operating system and the Gazebo simulation platform by conducting experiments. The results demonstrate that the APF-D3QNPER algorithm exhibits excellent generalization abilities in the simulation environment, and the convergence speed, the loss value, the path planning time, and the path planning length of the APF-D3QNPER algorithm are all less than for other algorithms in diverse scenarios.
科研通智能强力驱动
Strongly Powered by AbleSci AI