强化学习
适应性
运动规划
计算机科学
局部最优
路径(计算)
趋同(经济学)
数学优化
避障
算法
收敛速度
人工智能
障碍物
机制(生物学)
任务(项目管理)
功能(生物学)
机器人
自适应采样
点(几何)
自适应算法
自适应系统
作者
Qian Xiao,Tin-Su Pan,Kexin Wang,Shan Cui
出处
期刊:Sensors
[Multidisciplinary Digital Publishing Institute]
日期:2025-07-29
卷期号:25 (15): 4685-4685
被引量:1
摘要
Traditional deep reinforcement learning methods suffer from slow convergence speeds and poor adaptability in complex environments and are prone to falling into local optima in AGV system applications. To address these issues, in this paper, an adaptive path planning algorithm with an improved Deep Q Network algorithm called the B-PER DQN algorithm is proposed. Firstly, a dynamic temperature adjustment mechanism is constructed, and the temperature parameters in the Boltzmann strategy are adaptively adjusted by analyzing the change trend of the recent reward window. Next, the Priority experience replay mechanism is introduced to improve the training efficiency and task diversity through experience grading sampling and random obstacle configuration. Then, a refined multi-objective reward function is designed, combined with direction guidance, step punishment, and end point reward, to effectively guide the agent in learning an efficient path. Our experimental results show that, compared with other algorithms, the improved algorithm proposed in this paper achieves a higher success rate and faster convergence in the same environment and represents an efficient and adaptive solution for reinforcement learning for path planning in complex environments.
科研通智能强力驱动
Strongly Powered by AbleSci AI