计算卸载
马尔可夫决策过程
计算机科学
强化学习
边缘计算
资源配置
数学优化
移动边缘计算
云计算
分布式计算
计算
GSM演进的增强数据速率
能源消耗
马尔可夫过程
人工智能
计算机网络
工程类
算法
统计
数学
电气工程
操作系统
作者
Bin Qiu,Yunxiao Wang,Hailin Xiao,Zhongshan Zhang
标识
DOI:10.1109/tits.2024.3391831
摘要
As a novel paradigm, Vehicular Edge Computing (VEC) can effectively support computation-intensive or delay-sensitive applications in the Internet of Vehicles era. Computation offloading and resource management strategies are key technologies that directly determine the system cost in VEC networks. However, due to vehicle mobility and stochastic arrival computation tasks, designing an optimal offloading and resource allocation policy is extremely challenging. To solve this issue, a deep reinforcement learning-based intelligent offloading and power allocation scheme is proposed for minimizing the total delay cost and energy consumption in dynamic heterogeneous VEC networks. Specifically, we first construct an end-edge-cloud offloading model in a bidirectional road scenario, taking into account stochastic task arrival, time-varying channel conditions, and vehicle mobility. With the objective of minimizing the long-term total cost composed of the energy consumption and task delay, the Markov Decision Process (MDP) can be employed to solve such optimization problems. Moreover, considering the high-dimensional continuity of the action space and the dynamics of task generation, we propose a deep deterministic policy gradient-based adaptive computation offloading and power allocation (DDPG-ACOPA) algorithm to solve the formulated MDP problem. Extensive simulation results demonstrate that the proposed DDPG-ACOPA algorithm performs better in the dynamic heterogeneous VEC environment, significantly outperforming the other four baseline schemes.
科研通智能强力驱动
Strongly Powered by AbleSci AI