强化学习
计算机科学
移动边缘计算
分布式计算
马尔可夫决策过程
边缘计算
计算卸载
资源配置
边缘设备
GSM演进的增强数据速率
云计算
资源管理(计算)
马尔可夫过程
人工智能
计算机网络
统计
操作系统
数学
作者
Ying Chen,Zhiyong Liu,Yongchao Zhang,Yuan Wu,Xin Chen,Lian Zhao
标识
DOI:10.1109/tii.2020.3028963
摘要
Nowadays, driven by the rapid development of smart mobile equipments and 5G network technologies, the application scenarios of Internet of Things (IoT) technology are becoming increasingly widespread. The integration of IoT and industrial manufacturing systems forms the industrial IoT (IIoT). Because of the limitation of resources, such as the computation unit and battery capacity in the IIoT equipments (IIEs), computation-intensive tasks need to be executed in the mobile edge computing (MEC) server. However, the dynamics and continuity of task generation lead to a severe challenge to the management of limited resources in IIoT. In this article, we investigate the dynamic resource management problem of joint power control and computing resource allocation for MEC in IIoT. In order to minimize the long-term average delay of the tasks, the original problem is transformed into a Markov decision process (MDP). Considering the dynamics and continuity of task generation, we propose a deep reinforcement learning-based dynamic resource management (DDRM) algorithm to solve the formulated MDP problem. Our DDRM algorithm exploits the deep deterministic policy gradient and can deal with the high-dimensional continuity of the action and state spaces. Extensive simulation results demonstrate that the DDRM can reduce the long-term average delay of the tasks effectively.
科研通智能强力驱动
Strongly Powered by AbleSci AI