植绒(纹理)
计算机科学
多智能体系统
人工智能
分布式计算
复合材料
材料科学
作者
Jian Xiao,Chongjun Huang,Guohui Yuan,Yaoting Wang,Honyu Jia,Zhuoran Wang
标识
DOI:10.1109/tii.2024.3378804
摘要
Improving the performance of flocking control policies in practical scenarios is of great value in promoting the practical application of multiagent flocking collaborative control algorithms. In this article, concerning the practicality of flocking algorithms in stochastic communication environments, we propose a model learning based multiagent flocking control algorithm. First, an agent motion model construction method based on sequential attention mechanisms is proposed to provide a more realistic agent motion model for environmental interaction. Considering the cooperation and equivalence of agents in the flocking task, a multiagent cooperative soft actor–critic (MACSAC) algorithm is proposed to optimize the control policy model. Then, a digital learning system for multiagent flocking collaborative control is constructed by combining the learned motion model with the MACSAC algorithm. Finally, we design a behavior reasoning (BR) model based on the prior control policy, and introduce the model into the MACSAC algorithm to infer the motion state of noncommunicating adjacent agents, which solves the problem of poor control policy caused by the information loss of observation state in stochastic communication environments. The experimental results indicate that the constructed digital learning system can effectively simulate the policy learning of multiagent flocking in actual environmental scenarios, and demonstrate that the designed BR model can effectively improve the performance of the MACSAC-based multiagent flocking collaborative control algorithm in stochastic communication environments.
科研通智能强力驱动
Strongly Powered by AbleSci AI