计算机科学
强化学习
马尔可夫决策过程
适应性
分布式计算
调度(生产过程)
移动边缘计算
边缘计算
任务(项目管理)
资源配置
资源管理(计算)
GSM演进的增强数据速率
人工智能
马尔可夫过程
计算机网络
数学优化
生态学
统计
数学
管理
经济
生物
作者
Cong Wang,Tianye Yao,T.E. Fan,Sancheng Peng,Changming Xu,Shui Yu
标识
DOI:10.1109/jiot.2023.3294535
摘要
Existing mobile edge computing (MEC) systems are facing the challenges of limited resources and highly dynamic network environments. How to allocate resources to maintain the efficiency and timeliness of data and tasks is still an open issue. To address this problem, we propose a novel framework for UAV-assisted MEC systems using federated multi-agent reinforcement learning. First, we formulate a joint optimization problem as a multi-agent Markov decision process by jointly minimizing the average age of information and maximizing the number of recent tasks. Second, we design a novel scheduling algorithm for online collaborative resources by adopting multiple agents to learn and make decisions in accordance with the overall interests through federal learning. Finally, an experience replay mechanism for the internal experience pool is introduced to further improve learning efficiency. Experimental results show that our proposed algorithm is superior to the recent typical reinforcement learning-based algorithms. It not only has higher efficiency in task processing and data freshness, but also has more stable performance and adaptability across diverse experimental conditions.
科研通智能强力驱动
Strongly Powered by AbleSci AI