强化学习
计算机科学
趋同(经济学)
编码(内存)
运动规划
代表(政治)
运动(物理)
人工智能
编码(集合论)
状态空间
机器学习
离线学习
算法
机器人
在线学习
集合(抽象数据类型)
数学
统计
政治
万维网
政治学
法学
经济
程序设计语言
经济增长
作者
Chengmin Zhou,Bingding Huang,Pasi Fränti
标识
DOI:10.1109/tnnls.2023.3247160
摘要
Indoor motion planning challenges researchers because of the high density and unpredictability of moving obstacles. Classical algorithms work well in the case of static obstacles but suffer from collisions in the case of dense and dynamic obstacles. Recent reinforcement learning (RL) algorithms provide safe solutions for multiagent robotic motion planning systems. However, these algorithms face challenges in convergence: slow convergence speed and suboptimal converged result. Inspired by RL and representation learning, we introduced the ALN-DSAC: a hybrid motion planning algorithm where attention-based long short-term memory (LSTM) and novel data replay combine with discrete soft actor–critic (SAC). First, we implemented a discrete SAC algorithm, which is the SAC in the setting of discrete action space. Second, we optimized existing distance-based LSTM encoding by attention-based encoding to improve the data quality. Third, we introduced a novel data replay method by combining the online learning and offline learning to improve the efficacy of data replay. The convergence of our ALN-DSAC outperforms that of the trainable state of the arts. Evaluations demonstrate that our algorithm achieves nearly 100% success with less time to reach the goal in motion planning tasks when compared to the state of the arts. The test code is available at https://github.com/CHUENGMINCHOU/ALN-DSAC.
科研通智能强力驱动
Strongly Powered by AbleSci AI