强化学习
运动规划
钢筋
计算机科学
路径(计算)
人工智能
工程类
机器人
计算机网络
结构工程
作者
Jiandong Liu,Wei Luo,Guoqing Zhang,Ruihao Li
出处
期刊:Machines
[Multidisciplinary Digital Publishing Institute]
日期:2025-02-18
卷期号:13 (2): 162-162
标识
DOI:10.3390/machines13020162
摘要
In this paper, an enhanced deep reinforcement learning approach is presented for unmanned aerial vehicles (UAVs) operating in dynamic and potentially hazardous environments. Initially, the capability to discern obstacles from visual data is achieved through the application of the Yolov8-StrongSort technique. Concurrently, a novel data storage system for deep Q-networks (DQN), named dynamic data memory (DDM), is introduced to hasten the learning process and convergence for UAVs. Furthermore, addressing the issue of UAVs’ paths veering too close to obstacles, a novel strategy employing an artificial potential field to adjust the reward function is introduced, which effectively guides the UAVs away from proximate obstacles. Rigorous simulation tests in an AirSim-based environment confirm the effectiveness of these methods. Compared to DQN, dueling DQN, M-DQN, improved Q-learning, DDM-DQN, EPF (enhanced potential field), APF-DQN, and L1-MBRL, our algorithm achieves the highest success rate of 77.67%, while also having the lowest average number of moving steps. Additionally, we conducted obstacle avoidance experiments with UAVs with different densities of obstacles. These tests highlight fast learning convergence and real-time obstacle detection and avoidance, ensuring successful achievement of the target.
科研通智能强力驱动
Strongly Powered by AbleSci AI