强化学习
运动规划
计算机科学
路径(计算)
人工智能
钢筋
工程类
机器人
计算机网络
结构工程
作者
Jin Tang,Yangang Liang,Kebo Li
出处
期刊:Drones
[MDPI AG]
日期:2024-02-09
卷期号:8 (2): 60-60
被引量:33
标识
DOI:10.3390/drones8020060
摘要
Traditional unmanned aerial vehicle path planning methods focus on addressing planning issues in static scenes, struggle to balance optimality and real-time performance, and are prone to local optima. In this paper, we propose an improved deep reinforcement learning approach for UAV path planning in dynamic scenarios. Firstly, we establish a task scenario including an obstacle assessment model and model the UAV’s path planning problem using the Markov Decision Process. We translate the MDP model into the framework of reinforcement learning and design the state space, action space, and reward function while incorporating heuristic rules into the action exploration policy. Secondly, we utilize the Q function approximation of an enhanced D3QN with a prioritized experience replay mechanism and design the algorithm’s network structure based on the TensorFlow framework. Through extensive training, we obtain reinforcement learning path planning policies for both static and dynamic scenes and innovatively employ a visualized action field to analyze their planning effectiveness. Simulations demonstrate that the proposed algorithm can accomplish UAV dynamic scene path planning tasks and outperforms classical methods such as A*, RRT, and DQN in terms of planning effectiveness.
科研通智能强力驱动
Strongly Powered by AbleSci AI